Federated learning (FL) allows multiple clients to collaboratively learn a globally shared model through cycles of model aggregation and local model training without the need to share data. In this paper, we comprehensively study a new problem named aggregation error (AE), arising from the model aggregation stage on a server, which is mainly induced by the heterogeneity of the client data. Due to the large discrepancies between local models, the accompanying large AE generally results in a slow convergence and an expected reduction of accuracy for FL. In order to reduce AE, we propose a novel federated learning framework from a Bayesian perspective, in which a multivariate Gaussian product mechanism is employed to aggregate the local models. It is worth noting that the product of Gaussians is still a Gaussian. This property allows us to directly aggregate local expectations and covariances in a definitely convex form, thereby greatly reducing the AE. Accordingly, on the clients, we develop a new Federated Online Laplace Approximation (FOLA) method, which can estimate the parameters of the local posterior by repeatedly accumulating priors. Specifically, in every round, the global posterior distributed from the server can be treated as the priors, and thus the local posterior can also be effectively approximated by a Gaussian using FOLA. Experimental results on benchmarks reach state-of-the-arts performance and clearly demonstrate the advantages of the proposed method.
翻译:联邦学习(FL)让多个客户通过模型汇总和地方模型培训周期合作学习全球共享模式,无需分享数据。在本文中,我们全面研究服务器模型汇总阶段产生的名为聚合错误(AE)的新问题,这主要是客户数据的异质引起的。由于当地模型之间存在巨大差异,随之而来的大型AE通常导致FL的趋同速度缓慢和准确性预期下降。为了减少AE,我们从巴伊西亚的角度提出一个新的联合学习框架,其中采用多种变式高斯产品机制来汇总当地模型。值得注意的是,高斯人的产品仍然是高斯人的产品,这主要由于客户数据的异质性而引发的。这一属性使我们能够以绝对的 convex形式直接汇总当地的期望和共变异性,从而大大降低AE。 因此,在客户方面,我们开发了一个新的Federal-Opet Approximation (FOLA) 基准,该基准可以估计当地海报的参数,通过反复积累前方的图像和近似的图像分析结果,从而可以具体地从以往的服务器上展示。