We study collaborative machine learning systems where a massive dataset is distributed across independent workers which compute their local gradient estimates based on their own datasets. Workers send their estimates through a multipath fading multiple access channel with orthogonal frequency division multiplexing to mitigate the frequency selectivity of the channel. We assume that there is no channel state information (CSI) at the workers, and the parameter server (PS) employs multiple antennas to align the received signals. To reduce the power consumption and the hardware costs, we employ complex-valued low-resolution digital-to-analog converters (DACs) and analog-to-digital converters (ADCs), at the transmitter and the receiver sides, respectively, and study the effects of practical low-cost DACs and ADCs on the learning performance. Our theoretical analysis shows that the impairments caused by low-resolution DACs and ADCs, including those of one-bit DACs and ADCs, do not prevent the convergence of the federated learning algorithm, and the multipath channel effects vanish when a sufficient number of antennas are used at the PS. We also validate our theoretical results via simulations, and demonstrate that using low-resolution, even one-bit, DACs and ADCs causes only a slight decrease in the learning accuracy.
翻译:我们研究合作机器学习系统,在独立工人中间分配大量数据集,根据自己的数据集计算当地梯度估计数;工人通过多路径淡化的多访问频道发送估计数,使用正方频率分多路传输,以降低频道的频率选择性;我们假设工人没有频道状态信息,参数服务器(PS)使用多天线对接收信号进行协调;为减少电力消耗和硬件成本,我们在发报器和接收器两侧分别使用复杂价值的低分辨率数字对数值转换器和模拟对数字转换器(ADCs),并研究实际低成本发报机和ADC对学习绩效的影响。我们的理论分析显示,低分辨率发报和ADC(包括一位数的DACs和ADC)造成的缺陷,不会妨碍混合式学习算法的趋同,而且当PS使用足够数量的天线时多路径转换器和模拟数字转换器(ADCs)和模拟数字转换器(ADCs)就会消失。我们还通过模拟和微分辨率来验证我们的理论结果,我们只是通过模拟和低分辨率来证明我们的理论结果。