Adversarial attacks on deep-learning models have been receiving increased attention in recent years. Work in this area has mostly focused on gradient-based techniques, so-called white-box attacks, wherein the attacker has access to the targeted model's internal parameters; such an assumption is usually unrealistic in the real world. Some attacks additionally use the entire pixel space to fool a given model, which is neither practical nor physical (i.e., real-world). On the contrary, we propose herein a gradient-free method that uses the learned image manifold of a pretrained generative adversarial network (GAN) to generate naturalistic physical adversarial patches for object detectors. We show that our proposed method works both digitally and physically.
翻译:近年来,对深层学习模式的反向攻击日益受到重视,这一领域的工作主要侧重于基于梯度的技术,即所谓的白箱攻击,攻击者可在此获得定向模型的内部参数;这种假设在现实世界中通常不切实际。有些攻击还利用整个像素空间愚弄一个既不实际也不实际的(即现实世界)特定模型。相反,我们在此提出一种无梯度方法,利用预先训练的基因对抗网络(GAN)的学习图像元件,为物体探测器产生自然物理对立补丁。我们表明,我们所提议的方法在数字和物理上都是有效的。</s>