Existing model poisoning attacks to federated learning assume that an attacker has access to a large fraction of compromised genuine clients. However, such assumption is not realistic in production federated learning systems that involve millions of clients. In this work, we propose the first Model Poisoning Attack based on Fake clients called MPAF. Specifically, we assume the attacker injects fake clients to a federated learning system and sends carefully crafted fake local model updates to the cloud server during training, such that the learnt global model has low accuracy for many indiscriminate test inputs. Towards this goal, our attack drags the global model towards an attacker-chosen base model that has low accuracy. Specifically, in each round of federated learning, the fake clients craft fake local model updates that point to the base model and scale them up to amplify their impact before sending them to the cloud server. Our experiments show that MPAF can significantly decrease the test accuracy of the global model, even if classical defenses and norm clipping are adopted, highlighting the need for more advanced defenses.
翻译:联邦化学习中的现有中毒袭击模型假定攻击者可以接触到大量受损的真正客户。 但是,这种假设在涉及数百万客户的生产联合学习系统中是不现实的。 在这项工作中,我们提出第一个基于假冒客户名为MPAF的中毒袭击模型。 具体地说,我们假设攻击者将假客户注入一个联邦化学习系统,并在培训期间向云层服务器发送精心制作的假本地模型更新,这样学到的全球模型对于许多不加区分的测试投入的准确性就很低。 为了实现这一目标,我们的攻击将全球模型拖向一个精确度较低的攻击者选择基准模型。具体地说,在每轮联邦化学习中,假客户伪造本地模型,指向基础模型,并在将其送入云层服务器之前将其扩大影响。我们的实验显示,即使采用经典防御和常规剪贴,MPAF也能大幅降低全球模型的测试准确性,从而突显出更先进的防御的必要性。