Machine learning models have demonstrated promising performance in many areas. However, the concerns that they can be biased against specific groups hinder their adoption in high-stake applications. Thus it is essential to ensure fairness in machine learning models. Most of the previous efforts require access to sensitive attributes for mitigating bias. Nonetheless, it is often infeasible to obtain large scale of data with sensitive attributes due to people's increasing awareness of privacy and the legal compliance. Therefore, an important research question is how to make fair predictions under privacy? In this paper, we study a novel problem on fair classification in a semi-private setting, where most of the sensitive attributes are private and only a small amount of clean sensitive attributes are available. To this end, we propose a novel framework FairSP that can first learn to correct the noisy sensitive attributes under privacy guarantee via exploiting the limited clean sensitive attributes. Then, it jointly models the corrected and clean data in an adversarial way for debiasing and prediction. Theoretical analysis shows that the proposed model can ensure fairness when most of the sensitive attributes are private. Experimental results on real-world datasets demonstrate the effectiveness of the proposed model for making fair predictions under privacy and maintaining high accuracy.
翻译:机器学习模式在许多领域表现出了有希望的绩效。然而,它们可能对特定群体有偏见的担忧在许多领域表现出了良好的表现。然而,它们可能对特定群体持有偏见的担忧阻碍了它们被高科技应用的采纳。因此,确保机器学习模式的公平性至关重要。以前的大部分努力都要求获得敏感属性,以减少偏见。然而,由于人们日益意识到隐私和法律合规性,往往无法获得大量具有敏感属性的数据。因此,一个重要的研究问题是如何在隐私下作出公平的预测?在本文件中,我们研究了在半私人环境中公平分类的新问题,因为大多数敏感属性都是私有的,只有少量的清洁敏感属性。为此,我们提出了一个新的框架FairSP,它首先能够学习如何通过利用有限的清洁敏感属性来纠正隐私保障下的噪音敏感属性。然后,它用对抗性的方式联合模拟经过更正和清洁的数据,以便降低偏见和预测。理论分析表明,在大多数敏感属性是私有的情形下,拟议的模型可以确保公平性。现实世界数据集的实验结果表明,在隐私下进行公正的预测,并保持高度精确性。