Machine learning models are becoming pervasive in high-stakes applications. Despite their clear benefits in terms of performance, the models could show bias against minority groups and result in fairness issues in a decision-making process, leading to severe negative impacts on the individuals and the society. In recent years, various techniques have been developed to mitigate the bias for machine learning models. Among them, in-processing methods have drawn increasing attention from the community, where fairness is directly taken into consideration during model design to induce intrinsically fair models and fundamentally mitigate fairness issues in outputs and representations. In this survey, we review the current progress of in-processing bias mitigation techniques. Based on where the fairness is achieved in the model, we categorize them into explicit and implicit methods, where the former directly incorporates fairness metrics in training objectives, and the latter focuses on refining latent representation learning. Finally, we conclude the survey with a discussion of the research challenges in this community to motivate future exploration.
翻译:尽管这些模型在业绩方面有明显的好处,但它们可以显示对少数群体的偏见,并在决策过程中导致公正问题,对个人和社会产生严重的负面影响。近年来,已经开发了各种技术来减少对机器学习模型的偏见。其中,处理方法已引起社区越来越多的注意,在模型设计过程中直接考虑到公平性,以产生内在公平的模型,从根本上减轻产出和表述中的公平问题。在这次调查中,我们审查了处理中减少偏见技术目前的进展。根据在模型中实现的公平性,我们将它们分为明确和隐含的方法,前者将公平性指标直接纳入培训目标,后者侧重于改进潜在的代表性学习。最后,我们在调查结束时讨论了这一社区的研究挑战,以激发未来的探索。