Collecting training data from untrusted sources exposes machine learning services to poisoning adversaries, who maliciously manipulate training data to degrade the model accuracy. When trained on offline datasets, poisoning adversaries have to inject the poisoned data in advance before training, and the order of feeding these poisoned batches into the model is stochastic. In contrast, practical systems are more usually trained/fine-tuned on sequentially captured real-time data, in which case poisoning adversaries could dynamically poison each data batch according to the current model state. In this paper, we focus on the real-time settings and propose a new attacking strategy, which affiliates an accumulative phase with poisoning attacks to secretly (i.e., without affecting accuracy) magnify the destructive effect of a (poisoned) trigger batch. By mimicking online learning and federated learning on CIFAR-10, we show that the model accuracy will significantly drop by a single update step on the trigger batch after the accumulative phase. Our work validates that a well-designed but straightforward attacking strategy can dramatically amplify the poisoning effects, with no need to explore complex techniques.
翻译:从不信任的来源收集培训数据,使机器学习服务暴露给毒害对手,后者恶意地操纵培训数据,以降低模型的准确性。当在离线数据集上接受培训时,毒害对手必须在培训前预先注入中毒数据,而将这些中毒的批量装入模型的顺序是随机的。相比之下,实际系统通常更经常地根据顺序采集的实时数据进行培训/调整,在这种情况下,毒害对手可以根据当前模型状态对每批数据进行动态毒害。在本文中,我们侧重于实时设置并提出新的攻击战略,将中毒袭击的累积阶段结合到秘密(即,不影响准确性)放大(中毒)触发组的破坏性效应。通过模拟在线学习和在CIFAR-10上进行联合学习,我们表明模型准确性将显著下降,在累积阶段后,在触发批量上只需更新一个步骤。我们的工作证实,设计完善但直接的攻击战略可以极大地放大中毒效应,无需探索复杂的技术。