In federated learning, multiple client devices jointly learn a machine learning model: each client device maintains a local model for its local training dataset, while a master device maintains a global model via aggregating the local models from the client devices. The machine learning community recently proposed several federated learning methods that were claimed to be robust against Byzantine failures (e.g., system failures, adversarial manipulations) of certain client devices. In this work, we perform the first systematic study on local model poisoning attacks to federated learning. We assume an attacker has compromised some client devices, and the attacker manipulates the local model parameters on the compromised client devices during the learning process such that the global model has a large testing error rate. We formulate our attacks as optimization problems and apply our attacks to four recent Byzantine-robust federated learning methods. Our empirical results on four real-world datasets show that our attacks can substantially increase the error rates of the models learnt by the federated learning methods that were claimed to be robust against Byzantine failures of some client devices. We generalize two defenses for data poisoning attacks to defend against our local model poisoning attacks. Our evaluation results show that one defense can effectively defend against our attacks in some cases, but the defenses are not effective enough in other cases, highlighting the need for new defenses against our local model poisoning attacks to federated learning.
翻译:在联合学习中,多个客户设备联合学习一个机器学习模式:每个客户设备都保留一个本地培训数据集的本地模型,而一个主设备则通过汇总客户设备中的本地模型来维护一个全球模型。机器学习社区最近提议了一些联合会学习方法,声称这些方法对拜占庭某些客户设备的失败(例如系统故障、对抗性操纵)非常有力。在这项工作中,我们对本地模式中毒袭击进行了首次系统研究,以进行联合学习。我们假定袭击者已经损坏了一些客户设备,攻击者在学习过程中操纵了受损客户设备上的本地模型参数,这样全球模型就有一个很大的测试错误率。我们把攻击设计成优化问题,并将我们的攻击应用到四个最近的拜占庭-罗盘联合学习方法。我们四个真实世界数据集的实验结果表明,我们的攻击可以大大增加本地模式学习方法所学模型的错误率,这些模型被认为对拜占庭某些客户设备的失败十分有力。我们将两种防数据中毒袭击的防御方法概括为防数据中毒袭击的强大防御系统,我们需要在本地模式袭击中进行足够的防御,在新防御中进行辩护案例中进行足够的防御。