Federated learning allows many devices to collaborate in the training of machine learning models. As in traditional machine learning, there is a growing concern that models trained with federated learning may exhibit disparate performance for different demographic groups. Existing solutions to measure and ensure equal model performance across groups require access to information about group membership, but this access is not always available or desirable, especially under the privacy aspirations of federated learning. We study the feasibility of measuring such performance disparities while protecting the privacy of the user's group membership and the federated model's performance on the user's data. Protecting both is essential for privacy, because they may be correlated, and thus learning one may reveal the other. On the other hand, from the utility perspective, the privacy-preserved data should maintain the correlation to ensure the ability to perform accurate measurements of the performance disparity. We achieve both of these goals by developing locally differentially private mechanisms that preserve the correlations between group membership and model performance. To analyze the effectiveness of the mechanisms, we bound their error in estimating the disparity when optimized for a given privacy budget, and validate these bounds on synthetic data. Our results show that the error rapidly decreases for realistic numbers of participating clients, demonstrating that, contrary to what prior work suggested, protecting the privacy of protected attributes is not necessarily in conflict with identifying disparities in the performance of federated models.
翻译:与传统机器学习一样,人们日益担心的是,经过联合学习培训的模型可能在不同人口群体中表现出不同的业绩。 衡量和确保不同群体之间平等示范业绩的现有解决办法要求获得关于群体成员的信息,但这种获取并不总是可用或可取的,特别是在联合会学习的私隐愿望下。我们研究衡量这种业绩差异的可行性,同时保护用户群体成员的隐私和联合会模式在用户数据上的性能。保护两者对于隐私至关重要,因为它们可能相互关联,从而可以发现另一个。另一方面,从实用角度来看,隐私保护数据应当保持相关性,以确保准确衡量业绩差异的能力。我们通过发展地方差别化的私人机制,维护群体成员与模型业绩之间的相互关系,实现这两个目标。为了分析机制的有效性,我们控制了这些机制在优化特定隐私预算时估计差异的错误,并验证了这些在合成数据上的界限。我们的结果表明,从实用角度看,隐私保护的数据应该保持相关性,而过去的模型则表明,快速缩小了参与率,从而必然保护了参与业绩的客户。