Federated learning (FL) is an emerging paradigm for training machine learning models using possibly private data available at edge devices. The distributed operation of FL gives rise to challenges that are not encountered in centralized machine learning, including the need to preserve the privacy of the local datasets, and the communication load due to the repeated exchange of updated models. These challenges are often tackled individually via techniques that induce some distortion on the updated models, e.g., local differential privacy (LDP) mechanisms and lossy compression. In this work we propose a method coined joint privacy enhancement and quantization (JoPEQ), which jointly implements lossy compression and privacy enhancement in FL settings. In particular, JoPEQ utilizes vector quantization based on random lattice, a universal compression technique whose byproduct distortion is statistically equivalent to additive noise. This distortion is leveraged to enhance privacy by augmenting the model updates with dedicated multivariate privacy preserving noise. We show that JoPEQ simultaneously quantizes data according to a required bit-rate while holding a desired privacy level, without notably affecting the utility of the learned model. This is shown via analytical LDP guarantees, distortion and convergence bounds derivation, and numerical studies. Finally, we empirically assert that JoPEQ demolishes common attacks known to exploit privacy leakage.
翻译:联邦学习(FL)是利用边缘设备可能提供的私人数据培训机器学习模型的新兴范例。FL的分布式操作在中央机器学习中产生了没有遇到的挑战,包括需要维护本地数据集的隐私,以及由于反复交换更新模型而产生的通信负荷。这些挑战往往通过导致对更新模型进行某些扭曲的技术,如地方差异隐私机制和损失压缩等,单独解决,在这项工作中,我们提议了一种创建联合隐私增强和量化的方法(JoPEQ),在FL设置中联合实施损失压缩和隐私增强。特别是,JoPEQ利用基于随机拉蒂的矢量量化,即产品扭曲在统计上等同于添加噪音的一种通用压缩技术。这种扭曲通过专门使用多变式隐私保护噪音来增强模型更新来增强隐私。我们表明,JoPEQ同时按照要求的比特率对数据进行量化,同时保持理想的隐私水平,同时不明显地影响所学模型的效用。这通过LDP保证、扭曲和趋同的我们所了解的隐私研究,最终通过分析性实验性研究来摧毁我们已知的隐私升级。