In this paper, we study the challenging task of Byzantine-robust decentralized training on arbitrary communication graphs. Unlike federated learning where workers communicate through a server, workers in the decentralized environment can only talk to their neighbors, making it harder to reach consensus and benefit from collaborative training. To address these issues, we propose a ClippedGossip algorithm for Byzantine-robust consensus and optimization, which is the first to provably converge to a $O(\delta_{\max}\zeta^2/\gamma^2)$ neighborhood of the stationary point for non-convex objectives under standard assumptions. Finally, we demonstrate the encouraging empirical performance of ClippedGossip under a large number of attacks.
翻译:在本文中,我们研究了在任意通信图上进行拜占庭容错分布式训练的挑战性任务。与工人通过服务器进行通信的联邦学习不同,去中心化环境中的工人只能与其邻居交流,这使得达成共识和从协作训练中受益变得更加困难。为了解决这些问题,我们提出了一种ClippedGossip算法,用于拜占庭容错的共识和优化,这是第一个在标准假设下的非凸目标函数下收敛于$O (\delta_{\max} \zeta^2/ \gamma^2)$稳定点邻域的算法。最后,我们证明了在大量攻击下ClippedGossip算法的鼓舞人心的实证性能。