The Private Aggregation of Teacher Ensembles (PATE) is an important private machine learning framework. It combines multiple learning models used as teachers for a student model that learns to predict an output chosen by noisy voting among the teachers. The resulting model satisfies differential privacy and has been shown effective in learning high-quality private models in semisupervised settings or when one wishes to protect the data labels. This paper asks whether this privacy-preserving framework introduces or exacerbates bias and unfairness and shows that PATE can introduce accuracy disparity among individuals and groups of individuals. The paper analyzes which algorithmic and data properties are responsible for the disproportionate impacts, why these aspects are affecting different groups disproportionately, and proposes guidelines to mitigate these effects. The proposed approach is evaluated on several datasets and settings.
翻译:私人集体教师组织(PATE)是一个重要的私人机器学习框架,它结合了作为学生模式教师的多种学习模式,学生模式学会预测教师通过吵闹的投票选择的产出。由此形成的模式满足了不同的隐私,在半监督环境中或在有人希望保护数据标签时,在学习高质量的私人模式方面显示出了有效性。本文询问这一隐私保护框架是否引入或加剧了偏向和不公平,并表明PATE可以引入个人和群体之间的准确性差异。论文分析了哪些算法和数据属性造成了不成比例的影响,为什么这些方面对不同群体的影响不成比例,并提出了减轻这些影响的指导方针。本文件对一些数据集和设置进行了评估。