Communication of model updates between client nodes and the central aggregating server is a major bottleneck in federated learning, especially in bandwidth-limited settings and high-dimensional models. Gradient quantization is an effective way of reducing the number of bits required to communicate each model update, albeit at the cost of having a higher error floor due to the higher variance of the stochastic gradients. In this work, we propose an adaptive quantization strategy called AdaQuantFL that aims to achieve communication efficiency as well as a low error floor by changing the number of quantization levels during the course of training. Experiments on training deep neural networks show that our method can converge in much fewer communicated bits as compared to fixed quantization level setups, with little or no impact on training and test accuracy.
翻译:客户节点和中央集成服务器之间的模式更新通信是联合学习的一大瓶颈,特别是在带宽限制的设置和高维模型中。 渐进量化是减少每个模式更新通信所需比特数的有效方法,尽管由于随机梯度差异较大而导致差差幅下限较高。 在这项工作中,我们提出了一个适应性量化战略,称为AdaQuantFL,目的是通过改变培训过程中的量化水平实现通信效率和低误差底数。 深神经网络培训实验显示,与固定量化水平配置相比,我们的方法可以聚集在少得多的沟通比特数上,对培训和测试准确性几乎没有影响或没有影响。