The advent of powerful prediction algorithms led to increased automation of high-stake decisions regarding the allocation of scarce resources such as government spending and welfare support. This automation bears the risk of perpetuating unwanted discrimination against vulnerable and historically disadvantaged groups. Research on algorithmic discrimination in computer science and other disciplines developed a plethora of fairness metrics to detect and correct discriminatory algorithms. Drawing on robust sociological and philosophical discourse on distributive justice, we identify the limitations and problematic implications of prominent fairness metrics. We show that metrics implementing equality of opportunity only apply when resource allocations are based on deservingness, but fail when allocations should reflect concerns about egalitarianism, sufficiency, and priority. We argue that by cleanly distinguishing between prediction tasks and decision tasks, research on fair machine learning could take better advantage of the rich literature on distributive justice.
翻译:强大的预测算法的出现导致在分配政府支出和福利支助等稀缺资源方面作出高度决策的自动化程度提高。这种自动化具有使弱势和历史上处于不利地位的群体长期遭受不想要的歧视的风险。关于计算机科学和其他学科的算法歧视的研究发展了大量的公平衡量标准,以发现和纠正歧视性算法。根据关于分配正义的强有力的社会学和哲学论述,我们确定突出的公平衡量标准的局限性和问题影响。我们表明,只有在资源分配以值得为根据时,才适用实现机会平等的指标,但在分配时不能反映对平等主义、充足性和优先地位的关切。我们说,通过明确区分预测任务和决定任务,关于公平机器学习的研究可以更好地利用关于分配正义的丰富文献。