This paper focuses on Bayesian inference in a federated learning context (FL). While several distributed MCMC algorithms have been proposed, few consider the specific limitations of FL such as communication bottlenecks and statistical heterogeneity. Recently, Federated Averaging Langevin Dynamics (FALD) was introduced, which extends the Federated Averaging algorithm to Bayesian inference. We obtain a novel tight non-asymptotic upper bound on the Wasserstein distance to the global posterior for FALD. This bound highlights the effects of statistical heterogeneity, which causes a drift in the local updates that negatively impacts convergence. We propose a new algorithm VR-FALD* that uses control variates to correct the client drift. We establish non-asymptotic bounds showing that VR-FALD* is not affected by statistical heterogeneity. Finally, we illustrate our results on several FL benchmarks for Bayesian inference.
翻译:本文侧重于联邦学习背景下的贝叶斯推论。 虽然已经提出了若干分布式的MCMC算法,但很少考虑FL的具体局限性,如通信瓶颈和统计异质性。最近,引入了Federal Average Langevin Dynamics(FALD),将联邦挥动算法延伸至Bayesian推论。我们获得了一个新的非抽取性紧性非抽取性上层连接瓦塞尔斯坦距离与FALD全球后方距离。这一捆绑突出统计异质性的影响,导致当地更新中的漂移,对趋同产生消极影响。我们提出了新的VR-FALD* 算法,使用控制变量纠正客户漂移。我们建立了非抽取性界限,表明VR-FALD* 不受统计异质性的影响。最后,我们介绍了我们关于Bayesian推论的若干FL基准的结果。