Federated learning (FL) allows multiple clients to collaboratively learn a globally shared model through cycles of model aggregation and local model training, without the need to share data. Most existing FL methods train local models separately on different clients, and then simply average their parameters to obtain a centralized model on the server side. However, these approaches generally suffer from large aggregation errors and severe local forgetting, which are particularly bad in heterogeneous data settings. To tackle these issues, in this paper, we propose a novel FL framework that uses online Laplace approximation to approximate posteriors on both the client and server side. On the server side, a multivariate Gaussian product mechanism is employed to construct and maximize a global posterior, largely reducing the aggregation errors induced by large discrepancies between local models. On the client side, a prior loss that uses the global posterior probabilistic parameters delivered from the server is designed to guide the local training. Binding such learning constraints from other clients enables our method to mitigate local forgetting. Finally, we achieve state-of-the-art results on several benchmarks, clearly demonstrating the advantages of the proposed method.
翻译:联邦学习(FL)允许多个客户通过模型汇总和地方模型培训周期合作学习全球共享模式,而无需分享数据。大多数现有的FL方法对不同的客户分别培训当地模型,然后只是平均其参数,以获得服务器方面的集中模型。然而,这些方法通常会遇到巨大的汇总错误和严重的本地遗忘,在多种数据设置中,这些错误和严重的本地遗忘尤其糟糕。为了解决这些问题,我们在本文件中提议了一个新型FL框架,使用在线拉普尔近似近似来近似客户和服务器方面的后遗星。在服务器方面,采用了多变量高斯产品机制来构建和最大化全球后遗物,主要减少了由本地模型之间的巨大差异引发的汇总错误。在客户方面,使用服务器提供的全球外生概率参数的先前损失是为了指导当地培训。从其他客户中设置的这种学习限制使我们能够减少本地的遗忘。最后,我们在若干基准上取得了最新的结果,明确展示了拟议方法的优点。