Federated learning of deep learning models for supervised tasks, e.g. image classification and segmentation, has found many applications: for example in human-in-the-loop tasks such as film post-production where it enables sharing of domain expertise of human artists in an efficient and effective fashion. In many such applications, we need to protect the training data from being leaked when gradients are shared in the training process due to IP or privacy concerns. Recent works have demonstrated that it is possible to reconstruct the training data from gradients for an image-classification model when its architecture is known. However, there is still an incomplete theoretical understanding of the efficacy and failure of such attacks. In this paper, we analyse the source of training-data leakage from gradients. We formulate the problem of training data reconstruction as solving an optimisation problem iteratively for each layer. The layer-wise objective function is primarily defined by weights and gradients from the current layer as well as the output from the reconstruction of the subsequent layer, but it might also involve a 'pull-back' constraint from the preceding layer. Training data can be reconstructed when we solve the problem backward from the output of the network through each layer. Based on this formulation, we are able to attribute the potential leakage of the training data in a deep network to its architecture. We also propose a metric to measure the level of security of a deep learning model against gradient-based attacks on the training data.
翻译:对监督任务(如图像分类和分解)的深层次学习模式的联邦学习发现许多应用:例如,在诸如电影后制作等人际接触任务中,它能够以高效和有效的方式分享人类艺术家的域的专门知识。在许多此类应用中,我们需要保护培训数据,在培训过程中由于IP或隐私问题而共享梯度时,当梯度因IP或隐私问题而共享时,培训数据不会被泄漏。最近的工作表明,有可能从梯度中重建培训数据,以建立图像分类模型,但在了解其结构时,这种攻击的功效和失败还存在理论上的局限性。在本文件中,我们分析培训数据数据从梯度渗漏的来源。我们把培训数据重建的问题作为解决每个层的迭接问题的方法。从当前层的权重和梯度以及随后层的重建产出,主要界定了这些层次的目标功能,但也可能涉及前层的“反向”制约。当我们从深层中找出问题时,培训数据从梯度渗漏到从梯度渗漏到深层网络的构建时,培训数据水平也是我们从深层数据结构中学习的属性。