Gradient inversion attack enables recovery of training samples from model updates in federated learning (FL) and constitutes a serious threat to data privacy. To mitigate this vulnerability, prior work proposed both principled defenses based on differential privacy, as well as heuristic defenses based on gradient compression as countermeasures. These defenses have so far been very effective, in particular those based on gradient compression that allow the model to maintain high accuracy while greatly reducing the attack's effectiveness. In this work, we argue that such findings do not accurately reflect the privacy risk in FL, and show that existing defenses can be broken by a simple adaptive attack that trains a model using auxiliary data to learn how to invert gradients on both vision and language tasks.
翻译:为了减轻这种脆弱性,先前的工作提出了基于不同隐私原则的防御,以及基于梯度压缩的防御。 到目前为止,这些防御非常有效,特别是基于梯度压缩的防御,使模型能够保持高精确度,同时大大降低攻击的效能。在这项工作中,我们争辩说,这些发现没有准确地反映FL的隐私风险,并表明现有的防御可以通过简单的适应性攻击来打破,这种攻击训练模型使用辅助数据来学习如何在视觉和语言任务上使梯度倒转。