Federated learning describes the distributed training of models across multiple clients while keeping the data private on-device. In this work, we view the server-orchestrated federated learning process as a hierarchical latent variable model where the server provides the parameters of a prior distribution over the client-specific model parameters. We show that with simple Gaussian priors and a hard version of the well known Expectation-Maximization (EM) algorithm, learning in such a model corresponds to FedAvg, the most popular algorithm for the federated learning setting. This perspective on FedAvg unifies several recent works in the field and opens up the possibility for extensions through different choices for the hierarchical model. Based on this view, we further propose a variant of the hierarchical model that employs prior distributions to promote sparsity. By similarly using the hard-EM algorithm for learning, we obtain FedSparse, a procedure that can learn sparse neural networks in the federated learning setting. FedSparse reduces communication costs from client to server and vice-versa, as well as the computational costs for inference with the sparsified network - both of which are of great practical importance in federated learning.
翻译:在这项工作中,我们把服务器-刻式联结学习过程视为一个等级潜伏变量模型,服务器在其中提供客户特定模型参数的先前分布参数。我们用简单的高西亚前科和众所周知的预期-最大化算法硬版来显示,在这种模型中学习与FedAvg相对应的FedAvg是联合学习设置最受欢迎的算法。关于FedAvg的这种观点使外地最近的一些工作更加明晰,通过对等级模型的不同选择打开扩展的可能性。基于这个观点,我们进一步提议一个等级模型的变式,利用先前的分布来促进松散。同样,我们通过使用硬式EM算法学习,获得FedSparse,这个程序可以在联邦化学习环境中学习稀薄的神经网络。FedSparse降低了从客户到服务器和反向用户的通信成本,并且通过对磁盘化网络进行计算成本的计算,两者都具有巨大的实用性学习重要性。