Federated learning has quickly gained popularity with its promises of increased user privacy and efficiency. Previous works have shown that federated gradient updates contain information that can be used to approximately recover user data in some situations. These previous attacks on user privacy have been limited in scope and do not scale to gradient updates aggregated over even a handful of data points, leaving some to conclude that data privacy is still intact for realistic training regimes. In this work, we introduce a new threat model based on minimal but malicious modifications of the shared model architecture which enable the server to directly obtain a verbatim copy of user data from gradient updates without solving difficult inverse problems. Even user data aggregated over large batches -- where previous methods fail to extract meaningful content -- can be reconstructed by these minimally modified models.
翻译:联邦学习由于承诺提高用户隐私和效率而迅速获得受欢迎程度。以前的工作表明,联谊梯度更新包含的信息在某些情况下可以用来大致恢复用户数据。以前对用户隐私的这些攻击范围有限,甚至没有超过少数几个数据点的梯度更新,甚至没有超过几个数据点,使得一些人得出结论,数据隐私对于现实的培训制度来说仍然完好无损。在这项工作中,我们引入了一个新的威胁模式,其依据是对共享模型架构进行微小但恶意的修改,使服务器能够直接从梯度更新中获取用户数据的逐字记录,而不会解决困难的反面问题。甚至用这些最微小的修改模式来重建大批(以前的方法无法提取有意义的内容)的用户数据。