Federated learning aims to train predictive models for data that is distributed across clients, under the orchestration of a server. However, participating clients typically each hold data from a different distribution, whereby predictive models with strong in-distribution generalization can fail catastrophically on unseen domains. In this work, we argue that in order to generalize better across non-i.i.d. clients, it is imperative to only learn correlations that are stable and invariant across domains. We propose FL Games, a game-theoretic framework for federated learning for learning causal features that are invariant across clients. While training to achieve the Nash equilibrium, the traditional best response strategy suffers from high-frequency oscillations. We demonstrate that FL Games effectively resolves this challenge and exhibits smooth performance curves. Further, FL Games scales well in the number of clients, requires significantly fewer communication rounds, and is agnostic to device heterogeneity. Through empirical evaluation, we demonstrate that FL Games achieves high out-of-distribution performance on various benchmarks.
翻译:联邦学习的目的是在服务器的协调下,为不同客户之间分布的数据培训预测模型;然而,参与的客户通常都持有不同分布的数据,因此,在分布上高度概括的预测模型可能会在无形领域灾难性地失灵。在这项工作中,我们争辩说,为了更好地在非i.i.d.客户之间推广,必须只学习稳定和不同领域的内在联系。我们提议FL运动会,这是学习不同客户之间不易的因果特征的混合学习的游戏理论框架。虽然实现纳什平衡的培训,但传统的最佳应对战略有高频振荡。我们证明,FL运动会有效地解决了这一挑战,并展示了平稳的性能曲线。此外,FL运动会在客户数量上的规模要大得多,需要大大减少交流回合,而且对于设置异质性是不可知性的。我们通过经验评估,证明FL运动会在不同基准上取得了高分流业绩。