There is growing interest in applying distributed machine learning to edge computing, forming federated edge learning. Federated edge learning faces non-i.i.d. and heterogeneous data, and the communication between edge workers, possibly through distant locations and with unstable wireless networks, is more costly than their local computational overhead. In this work, we propose DONE, a distributed approximate Newton-type algorithm with fast convergence rate for communication-efficient federated edge learning. First, with strongly convex and smooth loss functions, DONE approximates the Newton direction in a distributed manner using the classical Richardson iteration on each edge worker. Second, we prove that DONE has linear-quadratic convergence and analyze its communication complexities. Finally, the experimental results with non-i.i.d. and heterogeneous data show that DONE attains a comparable performance to the Newton's method. Notably, DONE requires fewer communication iterations compared to distributed gradient descent and outperforms DANE and FEDL, state-of-the-art approaches, in the case of non-quadratic loss functions.
翻译:使用分布式机器学习边际计算,形成联邦边际学习。联邦边际学习面临非i.d.d.和各种数据,边际工人之间的交流可能通过遥远的地点和不稳定的无线网络进行,比他们当地的计算管理费用更高。在这项工作中,我们建议DONE,一种分布式的大约牛顿型算法,具有通信效率联邦边际学习的快速趋同率。首先,由于具有很强的粘合和平稳的损失功能,DONE使用典型的Richardson迭代法,以分布式的方式接近牛顿方向。第二,我们证明DONE具有线性赤道趋同和分析其通信复杂性。最后,非赤道损失功能的实验结果和混杂数据显示,与Newton方法的类似性能。值得注意的是,在非赤道损失功能方面,DONE要求比分布式梯度脱落和超越DNE和FEDL的通信频率少。