This paper proposes a Decentralized Stochastic Gradient Descent (DSGD) algorithm to solve distributed machine-learning tasks over wirelessly-connected systems, without the coordination of a base station. It combines local stochastic gradient descent steps with a Non-Coherent Over-The-Air (NCOTA) consensus scheme at the receivers, that enables concurrent transmissions by leveraging the waveform superposition properties of the wireless channels. With NCOTA, local optimization signals are mapped to a mixture of orthogonal preamble sequences and transmitted concurrently over the wireless channel under half-duplex constraints. Consensus is estimated by non-coherently combining the received signals with the preamble sequences and mitigating the impact of noise and fading via a consensus stepsize. NCOTA-DSGD operates without channel state information (typically used in over-the-air computation schemes for channel inversion) and leverages the channel pathloss to mix signals, without explicit knowledge of the mixing weights (typically known in consensus-based optimization). It is shown that, with a suitable tuning of decreasing consensus and learning stepsizes, the error (measured as Euclidean distance) between the local and globally optimum models vanishes with rate $\mathcal O(k^{-1/4})$ after $k$ iterations. NCOTA-DSGD is evaluated numerically by solving an image classification task on the MNIST dataset, cast as a regularized cross-entropy loss minimization. Numerical results depict faster convergence vis-\`a-vis running time than implementations of the classical DSGD algorithm over digital and analog orthogonal channels, when the number of learning devices is large, under stringent delay constraints.
翻译:本文建议采用分散式软体梯度源( DSGD) 算法, 在无基站协调的情况下, 解决无线连接系统上分散的机器学习任务, 不通过基站协调。 它将本地的随机梯度梯度下降步骤与接收器的不一致性超过Air( NCOTA) 共识方案结合起来, 通过利用无线频道的波形叠加特性, 使同步传输成为可能。 有了 NCOTA, 本地优化信号被映射成一个交错序序序序列, 并在无线频道中同时传送。 共识平面序列在半翻转制约下, 由非一致式组合组合组合的机器教学任务, 将收到的信号与序言序列合并, 减轻噪音和通过协商一致的下降影响。 NAC- DSGD 运行的轨算法( 以OGNLA 缩略图的缩略图计算), 以ODOLA 缩略图的缩略图为缩略图。