Recently, data heterogeneity among the training datasets on the local clients (a.k.a., Non-IID data) has attracted intense interest in Federated Learning (FL), and many personalized federated learning methods have been proposed to handle it. However, the distribution shift between the training dataset and testing dataset on each client is never considered in FL, despite it being general in real-world scenarios. We notice that the distribution shift (a.k.a., out-of-distribution generalization) problem under Non-IID federated setting becomes rather challenging due to the entanglement between personalized and spurious information. To tackle the above problem, we elaborate a general dual-regularized learning framework to explore the personalized invariance, compared with the exsiting personalized federated learning methods which are regularized by a single baseline (usually the global model). Utilizing the personalized invariant features, the developed personalized models can efficiently exploit the most relevant information and meanwhile eliminate spurious information so as to enhance the out-of-distribution generalization performance for each client. Both the theoretical analysis on convergence and OOD generalization performance and the results of extensive experiments demonstrate the superiority of our method over the existing federated learning and invariant learning methods, in diverse out-of-distribution and Non-IID data cases.
翻译:最近,当地客户培训数据集(a.k.a.a.,非IID数据)的数据差异性引起了联邦学习联合会(FL)的浓厚兴趣,并提出了许多个性化的联邦学习方法来处理该问题,然而,FL从未考虑培训数据集和每个客户测试数据集之间的分布变化,尽管在现实世界的情景中,这种变化是普遍的。我们注意到,由于个人化和虚假信息之间的缠绕,在非IID组合环境下的分布变化(a.k.a.,分配范围外的概括化)问题变得相当具有挑战性。为了解决上述问题,我们制定了一个普遍的双重化学习框架,探索个人化差异,而个人化的联邦化学习方法通过单一基线(通常是全球模型)加以规范。我们发现,利用个人化的个性化特点,开发的个人化模型能够有效地利用最相关的信息,同时消除令人生动的信息,从而增强每个客户的分化二类通用性化业绩。为了解决上述问题,我们制定了一个普遍的双重性化学习框架的理论性化分析,并展示了我们广泛性化方法的超级学习结果。