In consequential decision-making applications, mitigating unwanted biases in machine learning models that yield systematic disadvantage to members of groups delineated by sensitive attributes such as race and gender is one key intervention to strive for equity. Focusing on demographic parity and equality of opportunity, in this paper we propose an algorithm that improves the fairness of a pre-trained classifier by simply dropping carefully selected training data points. We select instances based on their influence on the fairness metric of interest, computed using an infinitesimal jackknife-based approach. The dropping of training points is done in principle, but in practice does not require the model to be refit. Crucially, we find that such an intervention does not substantially reduce the predictive performance of the model but drastically improves the fairness metric. Through careful experiments, we evaluate the effectiveness of the proposed approach on diverse tasks and find that it consistently improves upon existing alternatives.
翻译:在相应的决策应用中,减少机械学习模式中那些对种族和性别等敏感属性所划分的群体成员造成系统性不利条件的不想要的偏见,是争取公平的关键干预措施之一。在本文件中,我们以人口均等和机会平等为重点,建议一种算法,通过简单地放弃仔细选择的培训数据点,提高受过训练的分类者的公平性。我们根据它们对公平利益衡量标准的影响,选择一些实例,采用无限的顶尖菜刀法计算。降低培训点是原则上的,但在实践中并不要求重新设置培训点。 很显然,我们发现这种干预不会大幅度降低模型的预测性能,而是极大地改进了公平性衡量标准。我们通过仔细的实验,评估了拟议方法在各种任务上的有效性,发现它不断改进现有的替代方法。