Federated Learning (FL) enables the multiple participating devices to collaboratively contribute to a global neural network model while keeping the training data locally. Unlike the centralized training setting, the non-IID and imbalanced (statistical heterogeneity) training data of FL is distributed in the federated network, which will increase the divergences between the local models and global model, further degrading performance. In this paper, we propose a novel clustered federated learning (CFL) framework FedGroup based on a similarity-based client clustering strategy, in which we 1) group the training of clients based on the similarities between the clients' optimize directions for high training performance; 2) reduce the complexity of client clustering algorithm by decomposing the high-dimension low-sample size (HDLSS) direction vectors. 3) implement a newcomer device cold start mechanism based on the auxiliary global model for framework scalability and practicality. FedGroup can achieve improvements by dividing joint optimization into groups of sub-optimization, and can be combined with FedProx, the state-of-the-art federated optimization algorithm. We evaluate FedGroup and FedGrouProx (combined with FedProx) on several open datasets. The experimental results show that our proposed frameworks significantly improving absolute test accuracy by +14.7% on FEMNIST compared to FedAvg, +5.4% on Sentiment140 compared to FedProx.
翻译:与中央培训环境不同,FL的非IID和不平衡(统计异质性)培训数据在联邦网络中分布,这将增加当地模式与全球模式之间的差异,进一步降低业绩。在本文件中,我们提议根据基于类似性的客户群组合战略建立一个新型的组合式联合学习框架FedGroup(CFL),其中我们1)根据客户对高培训业绩的最佳方向之间的相似性,将客户培训组合起来;2)通过解开高分化低缩规模(HDLSS)方向矢量来降低客户组合算法的复杂性。3)根据辅助性全球模式实施一个新的冷藏启动机制,促进框架的可缩放性和实用性。FedGroup可以通过将联合优化分为次级优化小组实现改进,并与FedProx组合组合组合组合,即最先进的最佳优化状态组合组合;2)通过解析高分解低缩缩缩缩缩缩缩算法,降低客户群和美联会的绝对值。