Training ML models which are fair across different demographic groups is of critical importance due to the increased integration of ML in crucial decision-making scenarios such as healthcare and recruitment. Federated learning has been viewed as a promising solution for collaboratively training machine learning models among multiple parties while maintaining the privacy of their local data. However, federated learning also poses new challenges in mitigating the potential bias against certain populations (e.g., demographic groups), as this typically requires centralized access to the sensitive information (e.g., race, gender) of each datapoint. Motivated by the importance and challenges of group fairness in federated learning, in this work, we propose FairFed, a novel algorithm for fairness-aware aggregation to enhance group fairness in federated learning. Our proposed approach is server-side and agnostic to the applied local debiasing thus allowing for flexible use of different local debiasing methods across clients. We evaluate FairFed empirically versus common baselines for fair ML and federated learning, and demonstrate that it provides fairer models particularly under highly heterogeneous data distributions across clients. We also demonstrate the benefits of FairFed in scenarios involving naturally distributed real-life data collected from different geographical locations or departments within an organization.
翻译:联邦学习被视为多方合作培训机器学习模式的一个有希望的解决办法,同时维护当地数据的隐私;然而,联邦学习也带来了新的挑战,有助于减少对某些人口群体(如人口群体)的潜在偏见,因为这通常要求集中获取每个数据点的敏感信息(例如种族、性别),由于在医疗保健和招聘等关键决策情景中更多地将最低生活水平纳入关键决策情景(如医疗保健和招聘),因此这些培训模式至关重要。在这项工作中,我们提出公平意识整合的新算法,即公平意识汇总的新算法,以提高群体在联邦学习中的公平性。我们提议的方法是服务器一侧,对当地应用的偏向性也提出了新的挑战,从而可以灵活地使用不同客户的不同地方分化方法。我们从经验角度评价公平学习和联合学习的共同基线,并表明公平框架提供了更公平的模型,特别是在客户之间高度分散的数据分配之下。我们还展示了公平生命组织在不同地域或不同地点内自然分布的数据的效益。