Differentially private federated learning has been intensively studied. The current works are mainly based on the \textit{curator model} or \textit{local model} of differential privacy. However, both of them have pros and cons. The curator model allows greater accuracy but requires a trusted analyzer. In the local model where users randomize local data before sending them to the analyzer, a trusted analyzer is not required, but the accuracy is limited. In this work, by leveraging the \textit{privacy amplification} effect in the recently proposed shuffle model of differential privacy, we achieve the best of two worlds, i.e., accuracy in the curator model and strong privacy without relying on any trusted party. We first propose an FL framework in the shuffle model and a simple protocol (SS-Simple) extended from existing work. We find that SS-Simple only provides an insufficient privacy amplification effect in FL since the dimension of the model parameter is quite large. To solve this challenge, we propose an enhanced protocol (SS-Double) to increase the privacy amplification effect by subsampling. Furthermore, for boosting the utility when the model size is greater than the user population, we propose an advanced protocol (SS-Topk) with gradient sparsification techniques. We also provide theoretical analysis and numerical evaluations of the privacy amplification of the proposed protocols. Experiments on real-world datasets validate that SS-Topk improves the testing accuracy by 60.7\% than the local model based FL. We highlight the observation that SS-Topk even can improve by 33.94\% accuracy than the curator model based FL without any trusted party. Compared with non-private FL, our protocol SS-Topk only lose 1.48\% accuracy under $(4.696, 10^{-5})$-DP.
翻译:私密友联谊学习已经深入研究。 目前的工作主要基于差异隐私的\ textit{ curador 模型} 或\ textit{ 本地模型} 。 但是, 两者都具有正反两种世界的精度和密度。 设计者模型允许更高的准确性, 但需要一个值得信任的分析器。 在本地模型中, 用户在将本地数据发送到分析器之前随机化本地数据, 不需要一个可信任的分析器, 但准确性有限 。 在这项工作中, 利用最近提议的 差异隐私的闪烁模型的精度, 我们实现了两个世界的最佳的精度, 即: 调制导模型的精度和强度的隐私。 我们发现, SS-S- Sloint 只能提供一个不充分的隐私添加效果, 因为模型的维度是相当大的。 为了解决这项挑战, 我们建议一个强化的协议( SS- Double) 在不依赖任何信任方的情况下, 将F- L MIL 的精度分析方法提升了 。