Deep learning based image recognition systems have been widely deployed on mobile devices in today's world. In recent studies, however, deep learning models are shown vulnerable to adversarial examples. One variant of adversarial examples, called adversarial patch, draws researchers' attention due to its strong attack abilities. Though adversarial patches achieve high attack success rates, they are easily being detected because of the visual inconsistency between the patches and the original images. Besides, it usually requires a large amount of data for adversarial patch generation in the literature, which is computationally expensive and time-consuming. To tackle these challenges, we propose an approach to generate inconspicuous adversarial patches with one single image. In our approach, we first decide the patch locations basing on the perceptual sensitivity of victim models, then produce adversarial patches in a coarse-to-fine way by utilizing multiple-scale generators and discriminators. The patches are encouraged to be consistent with the background images with adversarial training while preserving strong attack abilities. Our approach shows the strong attack abilities in white-box settings and the excellent transferability in black-box settings through extensive experiments on various models with different architectures and training methods. Compared to other adversarial patches, our adversarial patches hold the most negligible risks to be detected and can evade human observations, which is supported by the illustrations of saliency maps and results of user evaluations. Lastly, we show that our adversarial patches can be applied in the physical world.
翻译:在当今世界上,在移动设备上广泛安装了深层次学习的图像识别系统。然而,在最近的研究中,深层次学习模式很容易受到对抗性实例的影响。一个称为对抗性补丁的对抗性例子,由于攻击能力强而引起研究人员的注意。虽然对立性补丁取得了高攻击成功率,但由于补丁和原始图像之间的视觉不一致,很容易发现这些补丁系统。此外,通常需要大量数据,用于文献中的对抗性补丁生成,而这种补丁在计算上成本昂贵且耗时。为了应对这些挑战,我们建议了一种方法,用单一的物理图像生成不明显的对立性补丁补丁。在我们的方法中,我们首先根据受害者模型的感知性敏感度来决定补丁点,然后通过使用多种规模的生成器和偏差的图像来以粗略的方式产生对立性补丁补丁补丁补丁补丁。鼓励补丁的补丁与背景图像保持一致,同时保持很强的攻击性培训能力。我们的方法可以显示白箱环境中的强攻击能力,以及黑箱环境中的极易移动性补补补补补补补补补补补补补补补补补。我们用模型的广泛实验,通过不同的模型和训练方法可以支持对等的对立性地展示其他模型。