Federated learning is a private-by-design distributed learning paradigm where clients train local models on their own data before a central server aggregates their local updates to compute a global model. Depending on the aggregation method used, the local updates are either the gradients or the weights of local learning models. Recent reconstruction attacks apply a gradient inversion optimization on the gradient update of a single minibatch to reconstruct the private data used by clients during training. As the state-of-the-art reconstruction attacks solely focus on single update, realistic adversarial scenarios are overlooked, such as observation across multiple updates and updates trained from multiple mini-batches. A few studies consider a more challenging adversarial scenario where only model updates based on multiple mini-batches are observable, and resort to computationally expensive simulation to untangle the underlying samples for each local step. In this paper, we propose AGIC, a novel Approximate Gradient Inversion Attack that efficiently and effectively reconstructs images from both model or gradient updates, and across multiple epochs. In a nutshell, AGIC (i) approximates gradient updates of used training samples from model updates to avoid costly simulation procedures, (ii) leverages gradient/model updates collected from multiple epochs, and (iii) assigns increasing weights to layers with respect to the neural network structure for reconstruction quality. We extensively evaluate AGIC on three datasets, CIFAR-10, CIFAR-100 and ImageNet. Our results show that AGIC increases the peak signal-to-noise ratio (PSNR) by up to 50% compared to two representative state-of-the-art gradient inversion attacks. Furthermore, AGIC is faster than the state-of-the-art simulation based attack, e.g., it is 5x faster when attacking FedAvg with 8 local steps in between model updates.
翻译:联邦学习是一种私自设计分布式学习模式,客户在中央服务器汇总本地更新以计算全球模型之前,用自己的数据对本地模型进行本地模型培训。根据所使用的汇总方法,本地更新要么是梯度,要么是本地学习模型的权重。最近的重建袭击在单一微型批次的梯度更新上应用了梯度倒置优化,以重建客户在培训中使用的私人数据。由于最先进的重建袭击完全侧重于单一更新,现实的对抗情景被忽视,例如,在多个小型服务器汇总其本地更新以计算全球模型。一些研究考虑一种更具挑战性的对抗性情景,即仅对基于多个小型终端的模型更新进行50个模型更新,或者根据本地学习模型的权重值。在本文件中,我们建议AGIC,一个新型的“最接近性变异端攻击”,在模型或梯度更新中,以及多个直径直的图像。在州里,AIC(i)从模型更新到高价比重的AIC,从模型更新到不断升级的ARC。