Over the past few years, Federated Learning (FL) has become a popular distributed machine learning paradigm. FL involves a group of clients with decentralized data who collaborate to learn a common model under the coordination of a centralized server, with the goal of protecting clients' privacy by ensuring that local datasets never leave the clients and that the server only performs model aggregation. However, in realistic scenarios, the server may be able to collect a small amount of data that approximately mimics the population distribution and has stronger computational ability to perform the learning process. To address this, we focus on the hybrid FL framework in this paper. While previous hybrid FL work has shown that the alternative training of clients and server can increase convergence speed, it has focused on the scenario where clients fully participate and ignores the negative effect of partial participation. In this paper, we provide theoretical analysis of hybrid FL under clients' partial participation to validate that partial participation is the key constraint on convergence speed. We then propose a new algorithm called FedCLG, which investigates the two-fold role of the server in hybrid FL. Firstly, the server needs to process the training steps using its small amount of local datasets. Secondly, the server's calculated gradient needs to guide the participated clients' training and the server's aggregation. We validate our theoretical findings through numerical experiments, which show that our proposed method FedCLG outperforms state-of-the-art methods.
翻译:近年来,联邦学习(Federated Learning, FL)已成为一种流行的分布式机器学习范式。FL涉及一组拥有分散数据的客户端在中央服务器的协调下协作学习共同的模型,旨在通过确保本地数据集永远不会离开客户端,并且服务器仅执行模型聚合来保护客户端隐私。然而,在实际场景中,服务器可以收集少量近似于人口分布的数据,并具有更强的计算能力来执行学习过程。针对此,本文关注混合FL框架。虽然以往的混合FL工作已经表明,客户端与服务器的替代训练可以提高收敛速度,但其专注于客户端完全参与的情况,忽略了部分参与的负面影响。本文通过对客户端部分参与的混合FL的理论分析,验证了部分参与是收敛速度的关键限制。随后,提出了一种新的算法FedCLG,探讨了服务器在混合FL中的双重角色。首先,服务器需要使用其少量的本地数据集处理训练步骤。其次,服务器计算的梯度需要指导参与的客户端和服务器的聚合训练。通过数值实验验证了理论发现,结果表明我们的提出的FedCLG方法优于现有最先进的方法。