The distributed nature and privacy-preserving characteristics of federated learning make it prone to the threat of poisoning attacks, especially backdoor attacks, where the adversary implants backdoors to misguide the model on certain attacker-chosen sub-tasks. In this paper, we present a novel method ARIBA to accurately and robustly identify backdoor attacks in federated learning. By empirical study, we observe that backdoor attacks are discernible by the filters of CNN layers. Based on this finding, we employ unsupervised anomaly detection to evaluate the pre-processed filters and calculate an anomaly score for each client. We then identify the most suspicious clients according to their anomaly scores. Extensive experiments are conducted, which show that our method ARIBA can effectively and robustly defend against multiple state-of-the-art attacks without degrading model performance.
翻译:联合会学习的分布性质和隐私保护特点使其容易受到中毒攻击的威胁,特别是后门攻击,敌手植入后门错误地引导某些攻击者选择的子任务模式。在本文中,我们介绍了一种新颖的ARIBA方法,以准确和有力地识别联盟学习过程中的后门攻击。根据经验研究,我们观察到有线电视新闻网层的过滤器可以辨识到后门攻击。基于这一发现,我们使用未经监督的异常点探测方法来评估预处理的过滤器,计算每个客户的异常分数。然后,我们根据最可疑的客户的异常分数来识别这些客户。我们进行了广泛的实验,表明我们的ARIBA方法能够有效和有力地防御多种最先进的攻击,而没有贬低的模型性能。