Recent studies have proven that deep neural networks are vulnerable to backdoor attacks. Specifically, by mixing a small number of poisoned samples into the training set, the behavior of the trained model can be maliciously controlled. Existing attack methods construct such adversaries by randomly selecting some clean data from the benign set and then embedding a trigger into them. However, this selection strategy ignores the fact that each poisoned sample contributes inequally to the backdoor injection, which reduces the efficiency of poisoning. In this paper, we formulate improving the poisoned data efficiency by the selection as an optimization problem and propose a Filtering-and-Updating Strategy (FUS) to solve it. The experimental results on CIFAR-10 and ImageNet-10 indicate that the proposed method is effective: the same attack success rate can be achieved with only 47% to 75% of the poisoned sample volume compared to the random selection strategy. More importantly, the adversaries selected according to one setting can generalize well to other settings, exhibiting strong transferability. The prototype code of our method is now available at https://github.com/xpf/Data-Efficient-Backdoor-Attacks.
翻译:最近的研究证明,深层神经网络很容易受到后门攻击。 具体地说,通过将少量有毒样本混入训练组,受过训练的模型的行为可以受到恶意控制。 现有的攻击方法通过随机地从良型集中选择一些干净数据,然后在其中嵌入触发器来构造这些对手。 但是,这种选择战略忽略了这样一个事实,即每个中毒样本都对后门注射有同等作用,这降低了中毒效率。 在本文中,我们通过选择来提高中毒数据的效率,将其作为一个优化问题,并提出一个过滤和升级战略(FUS)来解决这个问题。 CIFAR-10和图像Net-10的实验结果表明,拟议的方法是有效的:与随机选择战略相比,只有47%至75%的中毒样本数量可以达到同样的攻击成功率。 更重要的是,根据一个环境选择的对手可以向其他环境广泛推广,显示强大的可转移性。 我们方法的原型代码现在可在https://github.com/xpf/Data-Efficent-Backdor-Atackstackackks。