Even though recent years have seen many attacks exposing severe vulnerabilities in federated learning (FL), a holistic understanding of what enables these attacks and how they can be mitigated effectively is still lacking. In this work we demystify the inner workings of existing targeted attacks. We provide new insights into why these attacks are possible and why a definitive solution to FL robustness is challenging. We show that the need for ML algorithms to memorize tail data has significant implications for FL integrity. This phenomenon has largely been studied in the context of privacy; our analysis sheds light on its implications for ML integrity. In addition, we show how constraints on client updates can effectively improve robustness. To incorporate these constraints into secure FL protocols, we design and develop RoFL, a new secure FL system that enables constraints to be expressed and enforced on high-dimensional encrypted model updates. In essence, RoFL augments existing secure FL aggregation protocols with zero-knowledge proofs. Due to the scale of FL, realizing these checks efficiently presents a paramount challenge. We introduce several optimizations at the ML layer that allow us to reduce the number of cryptographic checks needed while preserving the effectiveness of our defenses. We show that RoFL scales to the sizes of models used in real-world FL deployments.
翻译:尽管近年来已经看到许多攻击,暴露了联谊学习(FL)中的严重弱点,但是仍然缺乏对这些攻击的哪些原因和如何有效减轻这些攻击的全面理解。在这项工作中,我们解开现有有目标攻击的内部功能的神秘性。我们提供了新的洞察力,说明为什么这些攻击是可能的,为什么最终解决FL的稳健性是具有挑战性的。我们表明,ML算法对记住尾部数据的必要性对FL的完整性有着重大影响。这一现象大部分是在隐私范围内研究的;我们的分析揭示了这些攻击对ML完整性的影响。此外,我们展示了对客户更新的制约如何能够有效地提高稳健性。为了将这些限制纳入安全的FL协议,我们设计和开发了一个新的安全FL系统,即一个新的安全FL系统,使得能够以高维的加密模型更新来表达和强制实施限制。从本质上讲,LLL用零知识证据强化现有安全的FL组合协议的必要性。由于FL的规模,这些检查的高效率构成了一项重大挑战。我们在ML层中引入了几项优化,使我们得以减少部署的密码检查数量,同时显示我们所使用的FL的防御规模。