Federated learning provides an effective paradigm to jointly optimize a model benefited from rich distributed data while protecting data privacy. Nonetheless, the heterogeneity nature of distributed data makes it challenging to define and ensure fairness among local agents. For instance, it is intuitively "unfair" for agents with data of high quality to sacrifice their performance due to other agents with low quality data. Currently popular egalitarian and weighted equity-based fairness measures suffer from the aforementioned pitfall. In this work, we aim to formally represent this problem and address these fairness issues using concepts from co-operative game theory and social choice theory. We model the task of learning a shared predictor in the federated setting as a fair public decision making problem, and then define the notion of core-stable fairness: Given $N$ agents, there is no subset of agents $S$ that can benefit significantly by forming a coalition among themselves based on their utilities $U_N$ and $U_S$ (i.e., $\frac{|S|}{N} U_S \geq U_N$). Core-stable predictors are robust to low quality local data from some agents, and additionally they satisfy Proportionality and Pareto-optimality, two well sought-after fairness and efficiency notions within social choice. We then propose an efficient federated learning protocol CoreFed to optimize a core stable predictor. CoreFed determines a core-stable predictor when the loss functions of the agents are convex. CoreFed also determines approximate core-stable predictors when the loss functions are not convex, like smooth neural networks. We further show the existence of core-stable predictors in more general settings using Kakutani's fixed point theorem. Finally, we empirically validate our analysis on two real-world datasets, and we show that CoreFed achieves higher core-stability fairness than FedAvg while having similar accuracy.
翻译:联邦学习提供了一个有效的范例,以共同优化一个从丰富的分布式数据中受益的模式,同时保护数据隐私。然而,分布式数据的异质性质使得定义和确保当地代理人之间的公平性成为挑战。例如,对于数据质量高的代理人来说,这是直觉的“不公平”的,因为质量低的数据而牺牲其业绩。目前流行的基于平等和加权的基于公平性的措施受到上述缺陷的影响。在这项工作中,我们的目标是正式代表这一问题,并利用合作性游戏理论和社会选择理论的概念解决这些公平问题。我们模拟了在联合型游戏环境中学习一个共享预测器的任务,作为公平的公共决策问题,然后定义了核心公平性概念的概念:鉴于美元代理人,没有一组能通过以质量低的代理人牺牲其业绩来牺牲其业绩。目前,基于公共事业的美元和以美元为基础的基于公平的公平性公平性公平性措施。我们的目标是正式地代表这一问题,而当我们的核心投资者在使用一种稳定性成本的货币联盟内部数据时,核心预测值的准确性数据在两个质量低的当地选择性网络中也显示一种稳定的货币联盟内部的准确性功能。