Recent studies show that deep neural networks (DNNs) are vulnerable to backdoor attacks. A backdoor DNN model behaves normally with clean inputs, whereas outputs attacker's expected behaviors when the inputs contain a pre-defined pattern called a trigger. However, in some tasks, the attacker cannot know the exact target that shows his/her expected behavior, because the task may contain a large number of classes and the attacker does not have full access to know the semantic details of these classes. Thus, the attacker is willing to attack multiple suspected targets to achieve his/her purpose. In light of this, in this paper, we propose the M-to-N backdoor attack, a new attack paradigm that allows an attacker to launch a fuzzy attack by simultaneously attacking N suspected targets, and each of the N targets can be activated by any one of its M triggers. To achieve a better stealthiness, we randomly select M clean images from the training dataset as our triggers for each target. Since the triggers used in our attack have the same distribution as the clean images, the inputs poisoned by the triggers are difficult to be detected by the input-based defenses, and the backdoor models trained on the poisoned training dataset are also difficult to be detected by the model-based defenses. Thus, our attack is stealthier and has a higher probability of achieving the attack purpose by attacking multiple suspected targets simultaneously in contrast to prior backdoor attacks. Extensive experiments show that our attack is effective against different datasets with various models and achieves high attack success rates (e.g., 99.43% for attacking 2 targets and 98.23% for attacking 4 targets on the CIFAR-10 dataset) when poisoning only an extremely small portion of the training dataset (e.g., less than 2%). Besides, it is robust to pre-processing operations and can resist state-of-the-art defenses.
翻译:最近的研究显示,深心神经网络(DNN)很容易受到幕后攻击。后门DNN模式通常使用干净的投入,而产出攻击者在输入含有预定义的触发模式时的预期行为。然而,在某些任务中,攻击者无法知道显示其预期行为的确切目标,因为任务可能包含大量类别,攻击者无法完全了解这些类别的语义细节。因此,攻击者愿意攻击多个疑似目标,以达到他/她的目的。鉴于此,在本文件中,我们提议M-N后门攻击,而产出攻击者预期的行为则在输入含有预定义的触发模式时,称为触发器。但是在某些任务中,攻击者无法知道显示他/她的预期行为的确切目标,因为任务可能包含大量的类别,攻击者无法完全了解这些类别的语义细节细节。因此,攻击者可以随机从训练数据集中选择M干净的图像作为我们每个目标的触发器。由于我们攻击中使用的触发器与清洁的图像一样分布,触发器的毒性攻击目标在前方,攻击目标的频率在前方数据中也很难被检测到前方数据输入。在前方数据输入。在前方的概率中,因此很难辨测测测测测测到前的概率。在前的概率攻击中,攻击的概率值是用来测得的概率攻击目标。