Federated learning (FL) is vulnerable to model poisoning attacks, in which malicious clients corrupt the global model via sending manipulated model updates to the server. Existing defenses mainly rely on Byzantine-robust FL methods, which aim to learn an accurate global model even if some clients are malicious. However, they can only resist a small number of malicious clients in practice. It is still an open challenge how to defend against model poisoning attacks with a large number of malicious clients. Our FLDetector addresses this challenge via detecting malicious clients. FLDetector aims to detect and remove the majority of the malicious clients such that a Byzantine-robust FL method can learn an accurate global model using the remaining clients. Our key observation is that, in model poisoning attacks, the model updates from a client in multiple iterations are inconsistent. Therefore, FLDetector detects malicious clients via checking their model-updates consistency. Roughly speaking, the server predicts a client's model update in each iteration based on its historical model updates using the Cauchy mean value theorem and L-BFGS, and flags a client as malicious if the received model update from the client and the predicted model update are inconsistent in multiple iterations. Our extensive experiments on three benchmark datasets show that FLDetector can accurately detect malicious clients in multiple state-of-the-art model poisoning attacks. After removing the detected malicious clients, existing Byzantine-robust FL methods can learn accurate global models.
翻译:联邦学习组织(FL)易受到模式中毒袭击的伤害,恶意客户通过向服务器发送经过操纵的模型更新而腐蚀全球模型。现有的防御主要依靠Byzantine-robust FL方法,这些方法旨在学习准确的全球模型,即使有些客户是恶意的。不过,他们实际上只能抵制少数恶意客户。它仍然是一个公开的挑战,如何与大量恶意客户一起防范模式中毒袭击。我们的FL探测器通过探测恶意客户来应对这一挑战。FL探测器的目的是检测和清除大多数恶意客户,例如Byzantine-robust FL方法能够利用剩余客户学习准确的全球模型。我们的主要观察是,在模式中毒袭击中,客户的模型更新是不一致的。因此,FL探测器通过检查其模型更新时发现恶意客户。粗略地说,服务器根据历史模型更新后,使用Causurentale orem and L-BFBGS(FIGS)方法来检测和将一个客户的准确模型定位为恶意的多重测试。我们从模型中不断更新的客户。