Over the past few years, the federated learning ($\texttt{FL}$) community has witnessed a proliferation of new $\texttt{FL}$ algorithms. However, our understating of the theory of $\texttt{FL}$ is still fragmented, and a thorough, formal comparison of these algorithms remains elusive. Motivated by this gap, we show that many of the existing $\texttt{FL}$ algorithms can be understood from an operator splitting point of view. This unification allows us to compare different algorithms with ease, to refine previous convergence results and to uncover new algorithmic variants. In particular, our analysis reveals the vital role played by the step size in $\texttt{FL}$ algorithms. The unification also leads to a streamlined and economic way to accelerate $\texttt{FL}$ algorithms, without incurring any communication overhead. We perform numerical experiments on both convex and nonconvex models to validate our findings.
翻译:在过去几年里,联合学习社区($\textt{FL}$)见证了新的$(textt{FL})算法的激增。然而,我们对$(textt{FL}$)理论的低估仍然支离破碎,对这些算法的彻底、正式的比较仍然难以实现。受这一差距的驱使,我们从操作者分裂的角度看可以理解现有的许多$(textt{FL}$)算法。这种统一使我们能够轻松比较不同的算法,完善以前的趋同结果,并发现新的算法变方。特别是,我们的分析揭示了以$(textt{FL}$)算法的步数大小所起的重要作用。这种统一还导致一种简化的经济方式来加速$(textt{FLL}$)算法,而不会引起任何通信费。我们用数字实验对等模型和非convex模型进行验证我们的调查结果。