Data poisoning attacks, in which a malicious adversary aims to influence a model by injecting "poisoned" data into the training process, have attracted significant recent attention. In this work, we take a closer look at existing poisoning attacks and connect them with old and new algorithms for solving sequential Stackelberg games. By choosing an appropriate loss function for the attacker and optimizing with algorithms that exploit second-order information, we design poisoning attacks that are effective on neural networks. We present efficient implementations that exploit modern auto-differentiation packages and allow simultaneous and coordinated generation of tens of thousands of poisoned points, in contrast to existing methods that generate poisoned points one by one. We further perform extensive experiments that empirically explore the effect of data poisoning attacks on deep neural networks.
翻译:数据中毒攻击,恶意对手的目的是通过将“受污染”数据注入培训过程来影响模型,在这种攻击中,最近引起了人们的极大关注。在这项工作中,我们更仔细地研究现有的中毒攻击事件,并将它们与新旧的算法联系起来,以解决连续的斯塔克伯格游戏。我们选择攻击者的适当损失功能,并利用利用二级信息的算法优化,我们设计了对神经网络有效的中毒攻击事件。我们展示了利用现代自动差别包的高效应用,并允许同时和协调地生成数万个中毒点,而现有的方法是逐个产生中毒点。我们还进行了广泛的实验,从经验上探索数据中毒攻击对深神经网络的影响。