This paper investigates the role of dimensionality reduction in efficient communication and differential privacy (DP) of the local datasets at the remote users for over-the-air computation (AirComp)-based federated learning (FL) model. More precisely, we consider the FL setting in which clients are prompted to train a machine learning model by simultaneous channel-aware and limited communications with a parameter server (PS) over a Gaussian multiple-access channel (GMAC), so that transmissions sum coherently at the PS globally aware of the channel coefficients. For this setting, an algorithm is proposed based on applying federated stochastic gradient descent (FedSGD) for training the minimum of a given loss function based on the local gradients, Johnson-Lindenstrauss (JL) random projection for reducing the dimension of the local updates, and artificial noise to further aid user's privacy. For this scheme, our results show that the local DP performance is mainly improved due to injecting noise of greater variance on each dimension while keeping the sensitivity of the projected vectors unchanged. This is while the convergence rate is slowed down compared to the case without dimensionality reduction. As the performance outweighs for the slower convergence, the trade-off between privacy and convergence is higher but is shown to lessen in high-dimensional regime yielding almost the same trade-off with much less communication cost.
翻译:本文调查了远程用户在超空计算(AirComp)基于联合学习(FL)模型中减少本地数据集在高效通信和不同隐私(DP)方面作用的维度作用。更准确地说,我们考虑了FL设置,即通过同步频道认知和与高山多进入频道(GMAC)的参数服务器(PS)进行有限的通信,促使客户培训机器学习模式,同时通过高山多进入频道(GMAC)与参数服务器(PS)进行通信,使PS全球对频道系数有更一致的认识,从而能够一致地进行传输。为此设定了一种算法,以采用基于超空计算(AirComp)基于联合梯度的梯度下降(FedSGD)模型(FedSGD)来培训一个最小的损失功能。更精确地说,我们考虑FL的设置,即客户通过同步的频道认知和有限的通信(PS)通过一个参数服务器(PS)在高空进入通道频道(GMAC)来培训一个机器学习模式。我们的结果显示,当地DP业绩的改善主要是由于在每一个层面的噪音造成更大的差异,同时使预测矢量的变化也保持了。 矢量的敏感度的敏感度下降。这与高度的趋慢化速度,而贸易的趋同比的递减慢,而贸易的趋同率则显示贸易的递减速度慢。