Federated learning (FL) is an emerging learning paradigm without violating users' privacy. However, large model size and frequent model aggregation cause serious communication bottleneck for FL. To reduce the communication volume, techniques such as model compression and quantization have been proposed. Besides the fixed-bit quantization, existing adaptive quantization schemes use ascending-trend quantization, where the quantization level increases with the training stages. In this paper, we first investigate the impact of quantization on model convergence, and show that the optimal quantization level is directly related to the range of the model updates. Given the model is supposed to converge with the progress of the training, the range of the model updates will gradually shrink, indicating that the quantization level should decrease with the training stages. Based on the theoretical analysis, a descending quantization scheme named FedDQ is proposed. Experimental results show that the proposed descending quantization scheme can save up to 65.2% of the communicated bit volume and up to 68% of the communication rounds, when compared with existing schemes.
翻译:联邦学习(FL)是一个新兴的学习模式,没有侵犯用户隐私。然而,大型模型规模和频繁的模型聚合会给FL造成严重的通信瓶颈。为了减少通信量,已经提出了模型压缩和量化等技术。除了固定位数的量化之外,现有的适应性量化计划还采用递升-trend量化,即量化水平随着培训阶段的增加而增加。在本文件中,我们首先调查量化对模型趋同的影响,并表明最佳量化水平与模型更新的范围直接相关。鉴于模型被认为与培训进展趋同,模型更新的范围将逐渐缩小,表明量化水平应随着培训阶段的减少。根据理论分析,提出了称为FedDQ的递减量化计划。实验结果显示,与现有计划相比,拟议的递减量化计划可以节省所通报的位数的65.2%,通信回合的68%。