Most works on the fairness of machine learning systems focus on the blind optimization of common fairness metrics, such as Demographic Parity and Equalized Odds. In this paper, we conduct a comparative study of several bias mitigation approaches to investigate their behaviors at a fine grain, the prediction level. Our objective is to characterize the differences between fair models obtained with different approaches. With comparable performances in fairness and accuracy, are the different bias mitigation approaches impacting a similar number of individuals? Do they mitigate bias in a similar way? Do they affect the same individuals when debiasing a model? Our findings show that bias mitigation approaches differ a lot in their strategies, both in the number of impacted individuals and the populations targeted. More surprisingly, we show these results even apply for several runs of the same mitigation approach. These findings raise questions about the limitations of the current group fairness metrics, as well as the arbitrariness, hence unfairness, of the whole debiasing process.
翻译:大多数关于机器学习系统的公平性的工作都侧重于对通用公平度量的盲点优化,如人口均等和偶数等。在本文中,我们对几种减少偏差的方法进行了比较研究,以调查其以细粒、预测水平等方法进行的行为。我们的目标是区分以不同方法获得的公平模型之间的差异。在公平和准确性方面,不同减少偏差的方法是否对类似人数的个人产生影响?它们是否以类似的方式减轻偏差?它们是否在贬低模型时对相同的个人产生影响?我们的研究结果表明,减少偏差的方法在战略上存在很大差异,在受影响的个人数量和目标人群方面都是如此。更令人惊讶的是,我们甚至将这些结果展示为同一减轻偏差方法的若干运行过程所应用。这些结果提出了当前群体公平度量的局限性,以及整个消除偏差过程的任意性,因而是不公正的。