In the emerging paradigm of federated learning (FL), large amount of clients, such as mobile devices, are used to train possibly high-dimensional models on their respective data. Due to the low bandwidth of mobile devices, decentralized optimization methods need to shift the computation burden from those clients to the computation server while preserving privacy and reasonable communication cost. In this paper, we focus on the training of deep, as in multilayered, neural networks, under the FL settings. We present Fed-LAMB, a novel federated learning method based on a layerwise and dimensionwise updates of the local models, alleviating the nonconvexity and the multilayered nature of the optimization task at hand. We provide a thorough finite-time convergence analysis for Fed-LAMB characterizing how fast its gradient decreases. We provide experimental results under iid and non-iid settings to corroborate not only our theory, but also exhibit the faster convergence of our method, compared to the state-of-the-art.
翻译:由于移动设备带宽度低,分散化优化方法需要将计算负担从这些客户转移到计算服务器,同时保护隐私和合理的通信成本。在本文中,我们的重点是在FL设置下对深层神经网络进行多层次培训。我们介绍了Fed-LAMB,这是基于对当地模型进行层次和层面更新的新颖的联邦化学习方法,它缓解了手头优化任务的不协调性和多层次性质。我们为Fed-LAMB提供了全面的有限时间趋同分析,说明其梯度下降速度有多快。我们在iid和非二位设置下提供实验结果,不仅证实了我们的理论,而且显示了我们方法与最新技术的更快趋同。