Federated Learning (FL) is a promising privacy-preserving distributed learning paradigm but suffers from high communication cost when training large-scale machine learning models. Sign-based methods, such as SignSGD \cite{bernstein2018signsgd}, have been proposed as a biased gradient compression technique for reducing the communication cost. However, sign-based algorithms could diverge under heterogeneous data, which thus motivated the development of advanced techniques, such as the error-feedback method and stochastic sign-based compression, to fix this issue. Nevertheless, these methods still suffer from slower convergence rates. Besides, none of them allows multiple local SGD updates like FedAvg \cite{mcmahan2017communication}. In this paper, we propose a novel noisy perturbation scheme with a general symmetric noise distribution for sign-based compression, which not only allows one to flexibly control the tradeoff between gradient bias and convergence performance, but also provides a unified viewpoint to existing stochastic sign-based methods. More importantly, the unified noisy perturbation scheme enables the development of the very first sign-based FedAvg algorithm ($z$-SignFedAvg) to accelerate the convergence. Theoretically, we show that $z$-SignFedAvg achieves a faster convergence rate than existing sign-based methods and, under the uniformly distributed noise, can enjoy the same convergence rate as its uncompressed counterpart. Extensive experiments are conducted to demonstrate that the $z$-SignFedAvg can achieve competitive empirical performance on real datasets and outperforms existing schemes.
翻译:联邦学习联合会(FL)是一个充满希望的隐私保护分布式学习模式,但在培训大型机器学习模式时,沟通成本仍然很高。基于信号的方法,如SignSGD\cite{bernstein2018ignsgd},被提议为一种有偏向的梯度压缩技术,以减少通信成本。然而,基于信号的算法可能会在不同的数据下出现差异,从而推动开发先进的技术,如错误反馈方法和基于随机压缩的信号压缩,以解决这个问题。然而,这些方法仍然在缓慢的趋同率下受到损害。此外,除了其中没有任何一种基于信号的方法,例如SignSGD(SignSG) (SignSG) {cite{cite{cmahan2017complace}) 被提议作为一种有偏向的斜度压缩技术。但是,基于信号的算法不仅可以灵活控制梯度偏向偏向偏向和趋同性信号的性交易,而且还可以统一现有基于信号的方法。更重要的是,统一的cloan rod roal ofbbbation (F) exalalal-alalalalalalal-alationalationalationalationaliz) lavelational ax) 正在显示以快速显示以我们正在正在显示以正正正正正正加速的Sxxxxxxxxxxxxxxxxxxxx。