Federated learning (FL) has rapidly risen in popularity due to its promise of privacy and efficiency. Previous works have exposed privacy vulnerabilities in the FL pipeline by recovering user data from gradient updates. However, existing attacks fail to address realistic settings because they either 1) require toy settings with very small batch sizes, or 2) require unrealistic and conspicuous architecture modifications. We introduce a new strategy that dramatically elevates existing attacks to operate on batches of arbitrarily large size, and without architectural modifications. Our model-agnostic strategy only requires modifications to the model parameters sent to the user, which is a realistic threat model in many scenarios. We demonstrate the strategy in challenging large-scale settings, obtaining high-fidelity data extraction in both cross-device and cross-silo federated learning.
翻译:联邦学习(FL)由于其对隐私和效率的许诺而迅速增加。以前的工作通过从梯度更新中恢复用户数据,暴露了FL管道中的隐私脆弱性。然而,现有的攻击未能解决现实环境,因为它们:(1) 需要小批量的玩具环境,或(2) 需要不切实际和明显的结构修改。我们引入了新战略,大幅提升现有攻击,对任意大体积和不进行建筑修改的批量进行操作。我们的示范不可知性战略只需要修改发给用户的模型参数,在许多情况下,这是一个现实的威胁模型。我们展示了挑战大规模环境的战略,在跨构件和跨筒联学习中获取高不洁的数据提取。