Federated learning is gaining popularity as it enables training of high-utility models across several clients without directly sharing their private data. As a downside, the federated setting makes the model vulnerable to various adversarial attacks in the presence of malicious clients. Specifically, an adversary can perform backdoor attacks to control model predictions via poisoning the training dataset with a trigger. In this work, we propose a mitigation for backdoor attacks in a federated learning setup. Our solution forces the model optimization trajectory to focus on the invariant directions that are generally useful for utility and avoid selecting directions that favor few and possibly malicious clients. Concretely, we consider the sign consistency of the pseudo-gradient (the client update) as an estimation of the invariance. Following this, our approach performs dimension-wise filtering to remove pseudo-gradient elements with low sign consistency. Then, a robust mean estimator eliminates outliers among the remaining dimensions. Our theoretical analysis further shows the necessity of the defense combination and illustrates how our proposed solution defends the federated learning model. Empirical results on three datasets with different modalities and varying number of clients show that our approach mitigates backdoor attacks with a negligible cost on the model utility.
翻译:联邦学习越来越受欢迎,因为它有助于在不直接分享私人数据的情况下对多个客户进行高实用模型的培训。 作为下坡,联邦环境使模型在恶意客户面前很容易受到各种对抗性攻击。 具体地说, 对手可以用触发器对培训数据集下毒毒, 来控制模型预测。 在这项工作中, 我们提议在联合学习的设置中减少后门攻击。 我们的解决方案迫使模型优化轨迹聚焦于通常对实用性有帮助的逆差方向, 避免选择有利于少数可能恶意客户的方向。 具体地说, 我们认为伪升级( 客户更新) 的标志一致性是变量的估计。 之后, 我们的方法会进行维度的过滤, 以低信号一致性的方式清除伪升级元素。 然后, 强势的中值估计器可以消除剩余维度的外线。 我们的理论分析进一步表明防御组合的必要性, 并表明我们提出的解决方案如何捍卫联邦学习模型。 三个数据设置模型( 客户更新) 的标志性结果以不同的方式和不同数量的公用率的客户的后门攻击, 显示我们可降低成本。