Federated learning performed by a decentralized networks of agents is becoming increasingly important with the prevalence of embedded software on autonomous devices. Bayesian approaches to learning benefit from offering more information as to the uncertainty of a random quantity, and Langevin and Hamiltonian methods are effective at realizing sampling from an uncertain distribution with large parameter dimensions. Such methods have only recently appeared in the decentralized setting, and either exclusively use stochastic gradient Langevin and Hamiltonian Monte Carlo approaches that require a diminishing stepsize to asymptotically sample from the posterior and are known in practice to characterize uncertainty less faithfully than constant step-size methods with a Metropolis adjustment, or assume strong convexity properties of the potential function. We present the first approach to incorporating constant stepsize Metropolis-adjusted HMC in the decentralized sampling framework, show theoretical guarantees for consensus and probability distance to the posterior stationary distribution, and demonstrate their effectiveness numerically on standard real world problems, including decentralized learning of neural networks which is known to be highly non-convex.
翻译:随着自动装置上嵌入式软件的普及,分权代理人网络开展的联邦学习变得越来越重要。贝伊斯学习方法从提供更多关于随机数量的不确定性的信息中受益,兰杰文和汉密尔顿方法有效地从具有较大参数的不确定分布中实现抽样;这种方法直到最近才出现在分权环境中,或者完全使用随机梯度的兰格文和汉密尔顿蒙特卡洛方法,这些方法要求逐步从后方逐步从后方进行无序抽样,而且在实践中人们知道,这些方法对不确定性的描述不如对大都会调整的固定级尺方法的忠实描述,或者对潜在功能具有很强的共性特性。我们介绍了在分权抽样框架中纳入不断逐步调整大都会调整的HMC的首个方法,展示了与后方静止分布的共识和概率距离的理论保障,并用数字方式展示了它们在标准的实际世界问题上的有效性,包括分散学习已知高度非连接的神经网络。