Federated learning frameworks have been regarded as a promising approach to break the dilemma between demands on privacy and the promise of learning from large collections of distributed data. Many such frameworks only ask collaborators to share their local update of a common model, i.e. gradients with respect to locally stored data, instead of exposing their raw data to other collaborators. However, recent optimization-based gradient attacks show that raw data can often be accurately recovered from gradients. It has been shown that minimizing the Euclidean distance between true gradients and those calculated from estimated data is often effective in fully recovering private data. However, there is a fundamental lack of theoretical understanding of how and when gradients can lead to unique recovery of original data. Our research fills this gap by providing a closed-form recursive procedure to recover data from gradients in deep neural networks. We name it Recursive Gradient Attack on Privacy (R-GAP). Experimental results demonstrate that R-GAP works as well as or even better than optimization-based approaches at a fraction of the computation under certain conditions. Additionally, we propose a Rank Analysis method, which can be used to estimate the risk of gradient attacks inherent in certain network architectures, regardless of whether an optimization-based or closed-form-recursive attack is used. Experimental results demonstrate the utility of the rank analysis towards improving the network's security. Source code is available for download from https://github.com/JunyiZhu-AI/R-GAP.
翻译:联邦学习框架被认为是打破对隐私的需求与从大量收集的分散数据中学习的希望之间的两难困境的一个很有希望的办法,许多这种框架只是要求合作者分享其当地对共同模型的更新,即当地储存数据的梯度,而不是将其原始数据暴露给其他合作者;然而,最近基于优化的梯度攻击表明,原始数据往往可以从梯度中准确恢复,实验结果表明,将真实梯度与估计数据计算出的梯度之间的欧西里德距离降到最低,往往能有效地完全恢复私人数据;然而,对于梯度如何和何时导致独特的原始数据恢复,根本缺乏理论上的理解。我们的研究填补了这一差距,提供了一种从深层神经网络中的梯度恢复数据的封闭式递归程序。我们称之为“对隐私的递增性攻击”(R-GAP)。实验结果显示,在一定条件下,将真实梯度梯度/精度计算法方法奏效,甚至比基于优化法的方法有效。此外,我们提议采用阶梯级分析法分析法分析方法,可以用来对深度攻击的系统化网络进行精确的升级,无论在何种程度上,还是在使用。