Although many fairness criteria have been proposed to ensure that machine learning algorithms do not exhibit or amplify our existing social biases, these algorithms are trained on datasets that can themselves be statistically biased. In this paper, we investigate the robustness of a number of existing (demographic) fairness criteria when the algorithm is trained on biased data. We consider two forms of dataset bias: errors by prior decision makers in the labeling process, and errors in measurement of the features of disadvantaged individuals. We analytically show that some constraints (such as Demographic Parity) can remain robust when facing certain statistical biases, while others (such as Equalized Odds) are significantly violated if trained on biased data. We also analyze the sensitivity of these criteria and the decision maker's utility to biases. We provide numerical experiments based on three real-world datasets (the FICO, Adult, and German credit score datasets) supporting our analytical findings. Our findings present an additional guideline for choosing among existing fairness criteria, or for proposing new criteria, when available datasets may be biased.
翻译:尽管已经提出了许多公平性标准以确保机器学习算法不会表现出或放大我们现有的社会偏见,但这些算法是在可能存在统计偏差的数据集上训练的。在本文中,我们调查了在算法在偏斜数据上训练时一些现有(人口统计学)公平性标准的稳健性。我们考虑数据集偏差的两种形式:标记误差和衡量不利地位个体特征的误差。我们从理论上证明了在面临某些统计偏差时,某些约束(如人口统计学平等)可以保持稳健,而其他约束(如等时机)则会在训练偏斜数据时显著违反。我们还分析了这些标准和决策者效用对偏差的敏感性。我们提供了基于三个真实世界数据集的数值实验,支持我们的理论发现。我们的研究结果为在数据集可能存在偏差的情况下选择现有公平性标准或提出新标准提供了额外的指导。