Federated learning (FL) is a promising strategy for performing privacy-preserving, distributed learning with a network of clients (i.e., edge devices). However, the data distribution among clients is often non-IID in nature, making efficient optimization difficult. To alleviate this issue, many FL algorithms focus on mitigating the effects of data heterogeneity across clients by introducing a variety of proximal terms, some incurring considerable compute and/or memory overheads, to restrain local updates with respect to the global model. Instead, we consider rethinking solutions to data heterogeneity in FL with a focus on local learning generality rather than proximal restriction. To this end, we first present a systematic study informed by second-order indicators to better understand algorithm effectiveness in FL. Interestingly, we find that standard regularization methods are surprisingly strong performers in mitigating data heterogeneity effects. Based on our findings, we further propose a simple and effective method, FedAlign, to overcome data heterogeneity and the pitfalls of previous methods. FedAlign achieves competitive accuracy with state-of-the-art FL methods across a variety of settings while minimizing computation and memory overhead. Code will be publicly available.
翻译:联邦学习(FL)是进行隐私保护、与客户网络(即边缘装置)进行分散学习的一个很有希望的战略。然而,客户之间数据分布的性质往往不是IID, 因而难以高效优化。为了缓解这一问题,许多FL算法侧重于通过采用各种精度术语,减轻数据在客户之间差异性的影响,有些涉及大量计算和(或)记忆管理,以限制全球模型方面的本地更新。相反,我们考虑重新思考FL数据差异性的解决办法,重点是本地学习通用性,而不是准目标限制。为此,我们首先根据二级指标进行系统研究,以更好地了解FL的算法效力。有趣的是,我们发现标准的正规化方法在减轻数据差异性影响方面表现出奇异。根据我们的研究结果,我们进一步提出一种简单有效的方法,即FedAlign,以克服数据差异性和以往方法的缺陷。FedAlign将获得具有竞争力的准确性,同时将可公开使用的存储器的代码用于最起码的版本。