The prevalent communication efficient federated learning (FL) frameworks usually take advantages of model gradient compression or model distillation. However, the unbalanced local data distributions (either in quantity or quality) of participating clients, contributing non-equivalently to the global model training, still pose a big challenge to these works. In this paper, we propose FedCliP, a novel communication efficient FL framework that allows faster model training, by adaptively learning which clients should remain active for further model training and pruning those who should be inactive with less potential contributions. We also introduce an alternative optimization method with a newly defined contribution score measure to facilitate active and inactive client determination. We empirically evaluate the communication efficiency of FL frameworks with extensive experiments on three benchmark datasets under both IID and non-IID settings. Numerical results demonstrate the outperformance of the porposed FedCliP framework over state-of-the-art FL frameworks, i.e., FedCliP can save 70% of communication overhead with only 0.2% accuracy loss on MNIST datasets, and save 50% and 15% of communication overheads with less than 1% accuracy loss on FMNIST and CIFAR-10 datasets, respectively.
翻译:常用的通信高效联合学习框架通常具有模型梯度压缩或模型蒸馏的优势,然而,参与客户的当地数据分布不平衡(数量或质量),不等同于全球模型培训,仍然给这些工作带来巨大的挑战。在本文中,我们提议FedCliP,这是一个新的通信高效FL框架,通过适应性学习,允许更快的模型培训,客户应积极参与进一步示范培训,并处理那些应该以较少的潜在贡献不活动的人。我们还采用另一种优化方法,采用新定义的缴款分数衡量标准,以便利积极和不活动客户的确定。我们用经验评估FL框架的通信效率,对ID和非IID设置下的三个基准数据集进行广泛试验。数字结果显示,Porposed FedCliP框架在最先进的FL框架方面表现超前,例如,FedCliP可以节省70%的通信间接费用,只损失0.2% MNIST数据集的准确度,并节省50%和15%的通信管理费,分别低于1 %的FMIS数据损失。