Federated learning has recently emerged as a paradigm promising the benefits of harnessing rich data from diverse sources to train high quality models, with the salient features that training datasets never leave local devices. Only model updates are locally computed and shared for aggregation to produce a global model. While federated learning greatly alleviates the privacy concerns as opposed to learning with centralized data, sharing model updates still poses privacy risks. In this paper, we present a system design which offers efficient protection of individual model updates throughout the learning procedure, allowing clients to only provide obscured model updates while a cloud server can still perform the aggregation. Our federated learning system first departs from prior works by supporting lightweight encryption and aggregation, and resilience against drop-out clients with no impact on their participation in future rounds. Meanwhile, prior work largely overlooks bandwidth efficiency optimization in the ciphertext domain and the support of security against an actively adversarial cloud server, which we also fully explore in this paper and provide effective and efficient mechanisms. Extensive experiments over several benchmark datasets (MNIST, CIFAR-10, and CelebA) show our system achieves accuracy comparable to the plaintext baseline, with practical performance.
翻译:最近,联邦学习成为一种范例,希望利用来自不同来源的丰富数据来培训高质量的模型,具有培训数据集从未离开当地装置的显著特点;只有模型更新是在当地计算和共享的,以便汇总产生一个全球模型;尽管联合会学习大大减轻隐私问题,而不是集中数据学习,但分享模型更新仍然构成隐私风险;在本文件中,我们提出了一个系统设计,在整个学习过程中有效地保护个人模型更新,允许客户仅提供模糊的模型更新,而云服务器仍然能够进行汇总。我们联合学习系统首先偏离先前的工作,支持轻量的加密和汇总,以及应对辍学客户的复原力,而不会影响他们参加未来几轮工作。与此同时,先前的工作主要忽略了密码领域的带宽效率优化,以及针对积极对抗性云端服务器的安保支持,我们也在本文件中充分探讨了这一点,并提供了有效和高效的机制。关于几个基准数据集(MNIST、CIFAR-10和CelibA)的广泛实验显示,我们的系统实现了可与平面基线相比的准确性,实际表现。