Federated learning aims at conducting inference when data are decentralised and locally stored on several clients, under two main constraints: data ownership and communication overhead. In this paper, we address these issues under the Bayesian paradigm. To this end, we propose a novel Markov chain Monte Carlo algorithm coined \texttt{QLSD} built upon quantised versions of stochastic gradient Langevin dynamics. To improve performance in a big data regime, we introduce variance-reduced alternatives of our methodology referred to as \texttt{QLSD}$^\star$ and \texttt{QLSD}$^{++}$. We provide both non-asymptotic and asymptotic convergence guarantees for the proposed algorithms and illustrate their benefits on several federated learning benchmarks.
翻译:联邦学习旨在根据两个主要制约因素,即数据所有权和通信管理费,对若干客户的数据分散和当地储存进行推断。本文在巴伊西亚范式下处理这些问题。为此,我们提议以四分五裂的随机梯度Langevin动态模型为基础,制作一部新颖的Markov链 Monte Carlo算法(conde phoned \ texttt\\ ⁇ LSD})。为了改进大数据系统的性能,我们引入了我们称为\ texttt ⁇ LSD}$和\ texttt ⁇ LSD}$的差别化替代方法。我们为拟议的算法提供非救济和无救济的趋同担保,并展示其在若干联邦学习基准上的好处。