Many applications of AI involve scoring individuals using a learned function of their attributes. These predictive risk scores are then used to take decisions based on whether the score exceeds a certain threshold, which may vary depending on the context. The level of delegation granted to such systems in critical applications like credit lending and medical diagnosis will heavily depend on how questions of fairness can be answered. In this paper, we study fairness for the problem of learning scoring functions from binary labeled data, a classic learning task known as bipartite ranking. We argue that the functional nature of the ROC curve, the gold standard measure of ranking accuracy in this context, leads to several ways of formulating fairness constraints. We introduce general families of fairness definitions based on the AUC and on ROC curves, and show that our ROC-based constraints can be instantiated such that classifiers obtained by thresholding the scoring function satisfy classification fairness for a desired range of thresholds. We establish generalization bounds for scoring functions learned under such constraints, design practical learning algorithms and show the relevance our approach with numerical experiments on real and synthetic data.
翻译:AI 的许多应用都涉及使用其属性的学习功能来评分个人。这些预测风险评分随后被用来根据评分是否超过某一阈值作出决定,这种阈值可能因背景不同而不同。在信用借贷和医学诊断等关键应用中,授予这类系统的授权程度在很大程度上取决于如何回答公平问题。在本文中,我们研究了从二进制标签数据中学习评分功能的公平问题,这是典型的学习任务,称为双方排名。我们争辩说,ROC曲线的功能性,即在此情况下排名准确的黄金标准衡量标准,可导致形成若干公平限制。我们采用了基于AC和ROC曲线的公平定义的一般组合,并表明我们基于ROC的限制可以即刻式地进行分类,这样通过评分功能的门槛获得的叙级人员就能够达到理想的阈值范围的公平性。我们为在这种制约下学到的评分功能规定了通用的界限,设计实用的学习算法,并表明我们的方法与实际和合成数据的数字实验的相关性。