Federated learning is considered as an effective privacy-preserving learning mechanism that separates the client's data and model training process. However, federated learning is still under the risk of privacy leakage because of the existence of attackers who deliberately conduct gradient leakage attacks to reconstruct the client data. Recently, popular strategies such as gradient perturbation methods and input encryption methods have been proposed to defend against gradient leakage attacks. Nevertheless, these defenses can either greatly sacrifice the model performance, or be evaded by more advanced attacks. In this paper, we propose a new defense method to protect the privacy of clients' data by learning to obscure data. Our defense method can generate synthetic samples that are totally distinct from the original samples, but they can also maximally preserve their predictive features and guarantee the model performance. Furthermore, our defense strategy makes the gradient leakage attack and its variants extremely difficult to reconstruct the client data. Through extensive experiments, we show that our proposed defense method obtains better privacy protection while preserving high accuracy compared with state-of-the-art methods.
翻译:联邦学习被视为一种有效的隐私保护学习机制,将客户的数据和模式培训过程区分开来。然而,联邦学习仍然处于隐私渗漏的风险之下,因为袭击者故意进行梯度渗漏攻击以重建客户数据。最近,人们提议了一些流行战略,例如梯度扰动方法和输入加密方法,以防范梯度渗漏攻击。然而,这些防御手段可以大大牺牲模型性能,或者被更先进的攻击所回避。我们在本文件中提出一种新的防御方法,通过学习模糊数据来保护客户数据的隐私。我们的防御方法可以产生与原始样本完全不同的合成样本,但它们也可以最大限度地保存其预测性能和保证模型性能。此外,我们的防御战略使得梯度渗漏攻击及其变种极难于重建客户数据。通过广泛的实验,我们证明我们提议的防御方法在保持与最新方法的高度精确性的同时,得到了更好的隐私保护。