Federated Learning (FL) has received increasing attention due to its privacy protection capability. However, the base algorithm FedAvg is vulnerable when it suffers from so-called backdoor attacks. Former researchers proposed several robust aggregation methods. Unfortunately, many of these aggregation methods are unable to defend against backdoor attacks. What's more, the attackers recently have proposed some hiding methods that further improve backdoor attacks' stealthiness, making all the existing robust aggregation methods fail. To tackle the threat of backdoor attacks, we propose a new aggregation method, X-raying Models with A Matrix (XMAM), to reveal the malicious local model updates submitted by the backdoor attackers. Since we observe that the output of the Softmax layer exhibits distinguishable patterns between malicious and benign updates, we focus on the Softmax layer's output in which the backdoor attackers are difficult to hide their malicious behavior. Specifically, like X-ray examinations, we investigate the local model updates by using a matrix as an input to get their Softmax layer's outputs. Then, we preclude updates whose outputs are abnormal by clustering. Without any training dataset in the server, the extensive evaluations show that our XMAM can effectively distinguish malicious local model updates from benign ones. For instance, when other methods fail to defend against the backdoor attacks at no more than 20% malicious clients, our method can tolerate 45% malicious clients in the black-box mode and about 30% in Projected Gradient Descent (PGD) mode. Besides, under adaptive attacks, the results demonstrate that XMAM can still complete the global model training task even when there are 40% malicious clients. Finally, we analyze our method's screening complexity, and the results show that XMAM is about 10-10000 times faster than the existing methods.
翻译:联邦学习组织(FL)因其隐私保护能力而日益受到越来越多的关注。然而,基础算法 FedAvg 在其遭受所谓的幕后攻击时很容易受到伤害。前研究人员提出了几种强大的聚合方法。不幸的是,许多这些聚合方法无法抵御幕后攻击。此外,攻击者最近提出了一些隐藏方法,以进一步改进幕后攻击的隐蔽性,使所有现有的强力聚合方法都失败。为了应对后门攻击的威胁,我们提议了一种新的汇总方法,即带有A 矩阵(XMAM)的X-直流模型,以披露幕后攻击者提交的恶意本地模型更新。由于我们观察Softmax层的输出显示在恶意和善意更新之间可辨别的模式。由于我们观察到Sftmax层的输出显示在恶意和良性更新之间可辨别的模式,我们后门攻击者在X-光检查时,我们用模型来调查本地模型更新的模型,在10-10 软性分析中,我们仍然可以排除其产出的反常态。在服务器上,不使用任何模拟培训数据更新的方法,在30次的服务器上,因此,最后的评估显示我们XMA没有方法。