Distributed data mining is an emerging research topic to effectively and efficiently address hard data mining tasks using big data, which are partitioned and computed on different worker nodes, instead of one centralized server. Nevertheless, distributed learning methods often suffer from the communication bottleneck when the network bandwidth is limited or the size of model is large. To solve this critical issue, many gradient compression methods have been proposed recently to reduce the communication cost for multiple optimization algorithms. However, the current applications of gradient compression to adaptive gradient method, which is widely adopted because of its excellent performance to train DNNs, do not achieve the same ideal compression rate or convergence rate as Sketched-SGD. To address this limitation, in this paper, we propose a class of novel distributed Adam-type algorithms (\emph{i.e.}, SketchedAMSGrad) utilizing sketching, which is a promising compression technique that reduces the communication cost from $O(d)$ to $O(\log(d))$ where $d$ is the parameter dimension. In our theoretical analysis, we prove that our new algorithm achieves a fast convergence rate of $O(\frac{1}{\sqrt{nT}} + \frac{1}{(k/d)^2 T})$ with the communication cost of $O(k \log(d))$ at each iteration. Compared with single-machine AMSGrad, our algorithm can achieve the linear speedup with respect to the number of workers $n$. The experimental results on training various DNNs in distributed paradigm validate the efficiency of our algorithms.
翻译:分布式数据挖掘是一个新兴的研究课题,目的是利用海量数据来切实有效地处理硬数据挖掘任务,而海量数据则是在不同的工人节点上进行分割和计算,而不是在一个集中的服务器上进行。然而,在网络带宽有限或模型大小大的情况下,分布式学习方法往往受到通信瓶颈的影响。为了解决这一关键问题,最近提出了许多梯度压缩方法,以降低多个优化算法的通信成本。然而,目前将梯度压缩应用于适应性梯度方法的应用,由于该方法在培训DNN的出色性能而被广泛采用,因此不会达到与Skecheched-SGD相同的理想压缩率或趋同率。为了应对这一限制,我们在本文件中建议使用一种新型分布式的亚当型算法(\i.e.},SketechedAMSGrad)使用草图,这是一个很有希望的压缩技术,可以将通信成本从$(d)美元降低到$(log)美元,因为其中的值是各种参数层面。在我们的理论分析中,我们的新算算方法实现了一个快速的以$(c_\\\\\\qral revalation) oration oration orage orage) orage orage orages $(ac) orgal real rec) orgal real recional real rec) $xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx