Decentralized federated learning (DFL) is a powerful framework of distributed machine learning and decentralized stochastic gradient descent (SGD) is a driving engine for DFL. The performance of decentralized SGD is jointly influenced by communication-efficiency and convergence rate. In this paper, we propose a general decentralized federated learning framework to strike a balance between communication-efficiency and convergence performance. The proposed framework performs both multiple local updates and multiple inter-node communications periodically, unifying traditional decentralized SGD methods. We establish strong convergence guarantees for the proposed DFL algorithm without the assumption of convex objective function. The balance of communication and computation rounds is essential to optimize decentralized federated learning under constrained communication and computation resources. For further improving communication-efficiency of DFL, compressed communication is applied to DFL, named DFL with compressed communication (C-DFL). The proposed C-DFL exhibits linear convergence for strongly convex objectives. Experiment results based on MNIST and CIFAR-10 datasets illustrate the superiority of DFL over traditional decentralized SGD methods and show that C-DFL further enhances communication-efficiency.
翻译:分散式联合学习(DFL)是分散式机器学习和分散式随机梯度下降(SGD)的强大框架,是DFL的驱动力。分散式SGD的表现受到通信效率和汇合率的共同影响。在本文件中,我们提议了一个普遍分散式联合学习框架,以平衡通信效率和汇合性业绩。拟议的框架定期进行多项地方更新和多节间交流,统一传统的分散式SGD方法。我们为拟议的DFL算法建立了强有力的趋同保证,而没有假定交错式客观功能。通信和计算周期的平衡对于在有限的通信和计算资源下优化分散式联合学习至关重要。为了进一步提高DFLL的通信效率,将压缩式通信应用到DFL,以压缩式通信命名DFL(C-DFL) 。拟议的C-DFL展示线性融合,以达到强烈的凝聚目标。基于MNIST和CIFAR-10数据集的实验结果说明DL优于传统的分散式SGDG方法,并显示C-DFL进一步加强通信效率。