The proliferation of Internet-of-Things (IoT) devices and cloud-computing applications over siloed data centers is motivating renewed interest in the collaborative training of a shared model by multiple individual clients via federated learning (FL). To improve the communication efficiency of FL implementations in wireless systems, recent works have proposed compression and dimension reduction mechanisms, along with digital and analog transmission schemes that account for channel noise, fading, and interference. This prior art has mainly focused on star topologies consisting of distributed clients and a central server. In contrast, this paper studies FL over wireless device-to-device (D2D) networks by providing theoretical insights into the performance of digital and analog implementations of decentralized stochastic gradient descent (DSGD). First, we introduce generic digital and analog wireless implementations of communication-efficient DSGD algorithms, leveraging random linear coding (RLC) for compression and over-the-air computation (AirComp) for simultaneous analog transmissions. Next, under the assumptions of convexity and connectivity, we provide convergence bounds for both implementations. The results demonstrate the dependence of the optimality gap on the connectivity and on the signal-to-noise ratio (SNR) levels in the network. The analysis is corroborated by experiments on an image-classification task.
翻译:为了提高无线系统实施FL的通信效率,最近的一些工作提出了压缩和尺寸减少机制,以及数字和模拟传输计划,其中考虑到频道噪音、消退和干扰; 先前的艺术主要侧重于分布式客户和一个中央服务器组成的恒星地形学; 与此形成对照的是,本文研究无线装置到构件网络的FL(D2D)网络的FL(FL),对分散型梯度梯度梯度下降的数码和模拟执行的绩效提供理论见解。 首先,我们引入通用数字和模拟无线执行通信效率DSGD算法,利用随机线性编码(RLC)进行压缩和超空计算; 下一步,根据调和连通性假设,我们为两个执行提供趋同点的FL(D2D)网络。