Unlike traditional central training, federated learning (FL) improves the performance of the global model by sharing and aggregating local models rather than local data to protect the users' privacy. Although this training approach appears secure, some research has demonstrated that an attacker can still recover private data based on the shared gradient information. This on-the-fly reconstruction attack deserves to be studied in depth because it can occur at any stage of training, whether at the beginning or at the end of model training; no relevant dataset is required and no additional models need to be trained. We break through some unrealistic assumptions and limitations to apply this reconstruction attack in a broader range of scenarios. We propose methods that can reconstruct the training data from shared gradients or weights, corresponding to the FedSGD and FedAvg usage scenarios, respectively. We propose a zero-shot approach to restore labels even if there are duplicate labels in the batch. We study the relationship between the label and image restoration. We find that image restoration fails even if there is only one incorrectly inferred label in the batch; we also find that when batch images have the same label, the corresponding image is restored as a fusion of that class of images. Our approaches are evaluated on classic image benchmarks, including CIFAR-10 and ImageNet. The batch size, image quality, and the adaptability of the label distribution of our approach exceed those of GradInversion, the state-of-the-art.
翻译:与传统的中央培训不同,联谊学习(FL)通过分享和汇总当地模型而不是当地数据来改善全球模型的绩效,从而通过分享和汇总当地模型而不是利用当地数据来保护用户隐私。虽然这种培训方法看起来是安全的,但一些研究显示,攻击者仍然可以根据共享梯度信息恢复私人数据。这种现场重建袭击值得深入研究,因为它可以在培训的任何阶段发生,无论是在培训开始时还是培训结束时发生;不需要相关的数据集,不需要再培训其他模型。我们打破了一些不切实际的假设和限制,以便在更广泛的情景中应用这次重建袭击。我们提出了可以将培训数据从共享梯度或重量(分别与FedSGD和FedAvg使用情景相对应)中重建的方法。我们建议了一种零点方法来恢复标签,即使批次中存在重复的标签;我们研究了标签与图像恢复之间的关系;我们发现即使批量中只有一个错误的标签;我们还发现,当批量图像带有同一标签时,对应的图像是恢复的,相应的图像是恢复的,包括10级图像的升级的升级,我们评估了10级图像的升级的等级。