Decentralized optimization is an emerging paradigm in distributed learning in which agents achieve network-wide solutions by peer-to-peer communication without the central server. Since communication tends to be slower than computation, when each agent communicates with only a few neighboring agents per iteration, they can complete iterations faster than with more agents or a central server. However, the total number of iterations to reach a network-wide solution is affected by the speed at which the agents' information is ``mixed'' by communication. We found that popular communication topologies either have large maximum degrees (such as stars and complete graphs) or are ineffective at mixing information (such as rings and grids). To address this problem, we propose a new family of topologies, EquiTopo, which has an (almost) constant degree and a network-size-independent consensus rate that is used to measure the mixing efficiency. In the proposed family, EquiStatic has a degree of $\Theta(\ln(n))$, where $n$ is the network size, and a series of time-dependent one-peer topologies, EquiDyn, has a constant degree of 1. We generate EquiDyn through a certain random sampling procedure. Both of them achieve an $n$-independent consensus rate. We apply them to decentralized SGD and decentralized gradient tracking and obtain faster communication and better convergence, theoretically and empirically. Our code is implemented through BlueFog and available at \url{https://github.com/kexinjinnn/EquiTopo}
翻译:分散化优化是分布式学习中的一种新兴范例,在这种模式中,代理商通过没有中央服务器的同行对等通信实现全网络范围的解决方案。由于通信往往比计算慢,当每个代理商与仅几个相邻的代理商通信时,他们完成迭代的速度可能比与更多代理商或中央服务器的迭代速度快。然而,实现全网络解决方案的迭代总数受到代理商信息通过通信“混合”的速度的影响。我们发现,流行的通信结构要么具有较大的最高程度(如星和完整图表),要么在混合信息(如环和网络)方面效率低下。为了解决这个问题,我们建议建立一个新的结构式组,即EquiTopopo,它具有一个(最基本)不变的程度,以及一个网络规模大小独立的共识率,用来测量混合效率。在拟议的大家庭中,EquiStatrial有一定的分流化(nx) 。在网络规模上是美元,在时间上依赖的一级和一级一级,在Equinal-deal-deal-deal rogrational aclutional a excial excial labal labal acal-deal acal acludeal acrecialmentalmental, 在Eyn slational acal acalmentalmentalmentalmentalmental axlation.