Federated learning is a machine learning training paradigm that enables clients to jointly train models without sharing their own localized data. However, the implementation of federated learning in practice still faces numerous challenges, such as the large communication overhead due to the repetitive server-client synchronization and the lack of adaptivity by SGD-based model updates. Despite that various methods have been proposed for reducing the communication cost by gradient compression or quantization, and the federated versions of adaptive optimizers such as FedAdam are proposed to add more adaptivity, the current federated learning framework still cannot solve the aforementioned challenges all at once. In this paper, we propose a novel communication-efficient adaptive federated learning method (FedCAMS) with theoretical convergence guarantees. We show that in the nonconvex stochastic optimization setting, our proposed FedCAMS achieves the same convergence rate of $O(\frac{1}{\sqrt{TKm}})$ as its non-compressed counterparts. Extensive experiments on various benchmarks verify our theoretical analysis.
翻译:联邦学习是一种机器学习训练范式,它使得客户端能够共同训练模型,而无需共享自己的本地数据。尽管采用了各种方法以减少由于重复的服务器-客户端同步而导致的通信开销,以及提出了FedAdam等联邦版本的自适应优化器来增加更多的适应性,但目前的联邦学习框架仍无法同时解决上述所有挑战。本文提出了一种新的通信效率自适应联邦学习方法(FedCAMS),并给出了理论收敛保证。我们展示了在非凸随机优化设置中,我们提出的FedCAMS达到了与其非压缩的对应物相同的收敛速度 $O(\frac{1}{\sqrt{TKm}})$。在各种基准测试中的大量实验验证了我们的理论分析。