The idea of federated learning is to train deep neural network models collaboratively and share them with multiple participants without exposing their private training data to each other. This is highly attractive in the medical domain due to patients' privacy records. However, a recently proposed method called Deep Leakage from Gradients enables attackers to reconstruct data from shared gradients. This study shows how easy it is to reconstruct images for different data initialization schemes and distance measures. We show how data and model architecture influence the optimal choice of initialization scheme and distance measure configurations when working with single images. We demonstrate that the choice of initialization scheme and distance measure can significantly increase convergence speed and quality. Furthermore, we find that the optimal attack configuration depends largely on the nature of the target image distribution and the complexity of the model architecture.
翻译:联盟式学习的理念是合作培训深神经网络模型,并与多个参与者分享这些模型,而不必相互披露其私人培训数据。由于病人的隐私记录,这在医疗领域具有很高的吸引力。然而,最近提出的名为“梯度的深渗漏”的方法使袭击者能够从共享梯度中重建数据。这项研究表明,重建不同数据初始化计划和距离测量的图像是多么容易。我们展示了数据和模型结构如何影响在使用单一图像时对初始化计划和远程测量配置的最佳选择。我们证明,初始化计划和远程测量的选择可以大大提高聚合速度和质量。此外,我们发现,最佳袭击配置在很大程度上取决于目标图像分布的性质和模型结构的复杂性。