Federated learning (FL) is a framework for users to jointly train a machine learning model. FL is promoted as a privacy-enhancing technology (PET) that provides data minimization: data never "leaves" personal devices and users share only model updates with a server (e.g., a company) coordinating the distributed training. We assess the realistic (i.e., worst-case) privacy guarantees that are provided to users who are unable to trust the server. To this end, we propose an attack against FL protected with distributed differential privacy (DDP) and secure aggregation (SA). The attack method is based on the introduction of Sybil devices that deviate from the protocol to expose individual users' data for reconstruction by the server. The underlying root cause for the vulnerability to our attack is the power imbalance. The server orchestrates the whole protocol and users are given little guarantees about the selection of other users participating in the protocol. Moving forward, we discuss requirements for an FL protocol to guarantee DDP without asking users to trust the server. We conclude that such systems are not yet practical.
翻译:联邦学习(FL)是用户联合培训机器学习模式的框架。FL是作为增强隐私的技术而推广的,可以提供尽量减少数据的最小化数据:数据从不“离开”个人设备,用户只与协调分布式培训的服务器(例如公司)共享模式更新。我们评估了向无法信任服务器的用户提供的现实(即最坏情况)隐私保障。为此,我们提议对受分布式不同隐私和安全聚合(SA)保护的FL进行攻击。攻击方法的基础是引入Sybil装置,这些装置偏离协议,暴露个人用户的数据,供服务器重建。易受我们攻击的根本原因是力量不平衡。服务器对参与协议的其他用户的选择几乎没有给予整个协议和用户的保障。向前看,我们讨论了在不要求用户信任服务器的情况下对FL协议进行保障DDP的要求。我们的结论是,这种系统尚不实用。