Machine learning algorithms are becoming integrated into more and more high-stakes decision-making processes, such as in social welfare issues. Due to the need of mitigating the potentially disparate impacts from algorithmic predictions, many approaches have been proposed in the emerging area of fair machine learning. However, the fundamental problem of characterizing Bayes-optimal classifiers under various group fairness constraints is not well understood as a theoretical benchmark. Based on the classical Neyman-Pearson argument (Neyman and Pearson, 1933; Shao, 2003) for optimal hypothesis testing, this paper provides a general framework for deriving Bayes-optimal classifiers under group fairness. This enables us to propose a group-based thresholding method that can directly control disparity, and more importantly, achieve an optimal fairness-accuracy tradeoff. These advantages are supported by experiments.
翻译:机器学习算法正逐渐被纳入越来越多的高层决策程序,如社会福利问题。由于需要减轻算法预测的潜在不同影响,在新兴的公平机器学习领域提出了许多办法。然而,在各种群体公平性限制下确定贝亚斯-最佳分类者为典型的理论标准这一根本问题并没有得到很好的理解。根据古典Neyman-Pearson论点(Neyman和Pearson,1933年;Shao,2003年)进行最佳假设测试,本文件为将贝亚斯-最佳分类者纳入群体公平性提供了一个总体框架。这使我们能够提出一个可以直接控制差异、更重要的是实现最佳公平-准确性权衡交易的集团门槛方法。这些优势得到了实验的支持。