Train machine learning models on sensitive user data has raised increasing privacy concerns in many areas. Federated learning is a popular approach for privacy protection that collects the local gradient information instead of real data. One way to achieve a strict privacy guarantee is to apply local differential privacy into federated learning. However, previous works do not give a practical solution due to three issues. First, the noisy data is close to its original value with high probability, increasing the risk of information exposure. Second, a large variance is introduced to the estimated average, causing poor accuracy. Last, the privacy budget explodes due to the high dimensionality of weights in deep learning models. In this paper, we proposed a novel design of local differential privacy mechanism for federated learning to address the abovementioned issues. It is capable of making the data more distinct from its original value and introducing lower variance. Moreover, the proposed mechanism bypasses the curse of dimensionality by splitting and shuffling model updates. A series of empirical evaluations on three commonly used datasets, MNIST, Fashion-MNIST and CIFAR-10, demonstrate that our solution can not only achieve superior deep learning performance but also provide a strong privacy guarantee at the same time.
翻译:关于敏感用户数据的机器学习模式在很多领域引起了越来越多的隐私问题。联邦学习是一种流行的隐私保护方法,收集当地的梯度信息,而不是真实数据。实现严格的隐私保障的一种方法是将本地差异隐私应用到联邦学习中。然而,由于三个问题,先前的工作并没有提供实际的解决办法。第一,噪音数据接近其原始价值,概率高,增加了信息暴露的风险。第二,对估计平均值造成很大差异,造成准确性差。最后,隐私预算由于深层学习模式中重量的高度而爆发。在本文件中,我们提议了一种新颖的本地差异隐私机制,用于联邦学习解决上述问题。它能够使数据与其原始价值更加不同,并带来更低的差异。此外,拟议的机制通过分裂和重新布置模型更新,绕过了维度的诅咒。对三种常用数据集,即MNIST、FASAshion-MNIST和CIFAR-10进行一系列经验评估,表明我们的解决办法不仅能够达到更高的深度学习表现,而且还提供了在同样时间提供强有力的隐私保障。