Federated learning has emerged as a popular technique for distributing model training across the network edge. Its learning architecture is conventionally a star topology between the devices and a central server. In this paper, we propose two timescale hybrid federated learning (TT-HF), which migrates to a more distributed topology via device-to-device (D2D) communications. In TT-HF, local model training occurs at devices via successive gradient iterations, and the synchronization process occurs at two timescales: (i) macro-scale, where global aggregations are carried out via device-server interactions, and (ii) micro-scale, where local aggregations are carried out via D2D cooperative consensus formation in different device clusters. Our theoretical analysis reveals how device, cluster, and network-level parameters affect the convergence of TT-HF, and leads to a set of conditions under which a convergence rate of O(1/t) is guaranteed. Experimental results demonstrate the improvements in convergence and utilization that can be obtained by TT-HF over state-of-the-art federated learning baselines.
翻译:联邦学习已成为在网络边缘传播示范培训的流行技术,其学习结构通常是一个装置和中央服务器之间的恒星地形。在本文中,我们提议两个时间尺度混合学习(TT-HF),通过设备对设备对设备(D2D)的通信迁移到分布更加分散的地形。在TT-HF中,地方模式培训通过连续的梯度迭代在设备上进行,同步进程在两个时间尺度上进行:(一)宏观尺度,通过设备-服务器互动进行全球聚合,和(二)微观尺度,通过D2D合作形成不同设备集群的共识进行局部集合。我们的理论分析揭示了设备、集群和网络层面参数如何影响TT-HF的趋同,并导致一系列条件,保证O(1/t)的趋同率。实验结果表明TT-HF超越了州级的联邦学习基线,在趋同和利用方面可以取得的改进。