Supervised learning models have been increasingly used for making decisions about individuals in applications such as hiring, lending, and college admission. These models may inherit pre-existing biases from training datasets and discriminate against protected attributes (e.g., race or gender). In addition to unfairness, privacy concerns also arise when the use of models reveals sensitive personal information. Among various privacy notions, differential privacy has become popular in recent years. In this work, we study the possibility of using a differentially private exponential mechanism as a post-processing step to improve both fairness and privacy of supervised learning models. Unlike many existing works, we consider a scenario where a supervised model is used to select a limited number of applicants as the number of available positions is limited. This assumption is well-suited for various scenarios, such as job application and college admission. We use ``equal opportunity'' as the fairness notion and show that the exponential mechanisms can make the decision-making process perfectly fair. Moreover, the experiments on real-world datasets show that the exponential mechanism can improve both privacy and fairness, with a slight decrease in accuracy compared to the model without post-processing.
翻译:在招聘、出借和大学录取等应用中,人们越来越多地使用受监督的学习模式来决定个人,这些模式可能继承培训数据集的原有偏见,并歧视受保护的属性(例如种族或性别),除了不公平外,在使用模型时还出现隐私问题,暴露了敏感的个人信息。在各种隐私概念中,差异隐私近年来变得很普遍。在这项工作中,我们研究使用差别化的私人指数机制作为后处理步骤的可能性,以提高受监督学习模式的公平和隐私。与许多现有工作不同,我们考虑一种设想,即使用受监督的模式来选择数量有限的申请人,因为现有的职位数量有限。这一假设适用于各种情况,例如职位申请和大学录取。我们利用“机会均等”作为公平概念,并表明指数机制可以使决策过程完全公平。此外,对现实世界数据集的实验表明,指数机制可以改善隐私和公平性,与没有后处理的模型相比,准确性略有下降。