The existence of real-world adversarial examples (commonly in the form of patches) poses a serious threat for the use of deep learning models in safety-critical computer vision tasks such as visual perception in autonomous driving. This paper presents an extensive evaluation of the robustness of semantic segmentation models when attacked with different types of adversarial patches, including digital, simulated, and physical ones. A novel loss function is proposed to improve the capabilities of attackers in inducing a misclassification of pixels. Also, a novel attack strategy is presented to improve the Expectation Over Transformation method for placing a patch in the scene. Finally, a state-of-the-art method for detecting adversarial patch is first extended to cope with semantic segmentation models, then improved to obtain real-time performance, and eventually evaluated in real-world scenarios. Experimental results reveal that, even though the adversarial effect is visible with both digital and real-world attacks, its impact is often spatially confined to areas of the image around the patch. This opens to further questions about the spatial robustness of real-time semantic segmentation models.
翻译:现实世界的对抗性实例(通常以补丁形式)的存在严重威胁了在安全关键计算机视觉任务中使用深层次学习模型,例如自动驾驶时的视觉感知;本文件对使用不同类型对抗性补丁(包括数字、模拟和物理补丁)攻击时的语义分解模型的稳健性进行了广泛评价;提出了一种新的损失功能,以提高攻击者诱使像素分类错误的能力;此外,还提出了一项新颖的攻击战略,以改进在现场放置补丁的 " 超变 " 方法;最后,首先扩展了探测对抗性补丁的先进方法,以适应语义分解模型,然后加以改进以获得实时性能,并最终在现实世界情景中加以评价。实验结果表明,即使数字式和现实世界攻击都可以看到对抗性效果,但其影响往往在空间上局限于补丁图周围的图像区域。这引起了关于实时分解模型空间稳健性的进一步问题。