Deep neural network (DNN) models have proven to be vulnerable to adversarial digital and physical attacks. In this paper, we propose a novel attack- and dataset-agnostic and real-time detector for both types of adversarial inputs to DNN-based perception systems. In particular, the proposed detector relies on the observation that adversarial images are sensitive to certain label-invariant transformations. Specifically, to determine if an image has been adversarially manipulated, the proposed detector checks if the output of the target classifier on a given input image changes significantly after feeding it a transformed version of the image under investigation. Moreover, we show that the proposed detector is computationally-light both at runtime and design-time which makes it suitable for real-time applications that may also involve large-scale image domains. To highlight this, we demonstrate the efficiency of the proposed detector on ImageNet, a task that is computationally challenging for the majority of relevant defenses, and on physically attacked traffic signs that may be encountered in real-time autonomy applications. Finally, we propose the first adversarial dataset, called AdvNet that includes both clean and physical traffic sign images. Our extensive comparative experiments on the MNIST, CIFAR10, ImageNet, and AdvNet datasets show that VisionGuard outperforms existing defenses in terms of scalability and detection performance. We have also evaluated the proposed detector on field test data obtained on a moving vehicle equipped with a perception-based DNN being under attack.
翻译:深心神经网络模型( DNN) 被证明容易受到对抗性数字和物理攻击。 在本文中, 我们提议为 DNN 的感知系统提供新型攻击和数据元件和实时检测器, 特别是, 提议的检测器依靠的观察是, 对抗性图像对某些标签变异性具有敏感性。 具体地说, 要确定图像是否受到对抗性操纵, 拟议的检测器检查是否在输入图像变换后对特定输入图像进行重大修改。 此外, 我们显示, 提议的检测器在运行和设计时都是计算性光, 从而适合实时应用 DNNNN 的感知系统。 为了突出这一点, 我们展示了图像网络上的拟议检测器的效率, 这个任务在计算上对大多数相关防御系统构成挑战, 在实时自主应用中可能遇到的受到实际攻击的交通信号。 最后, 我们提出了第一个对抗性数据集, 名为 AdvNet, 在运行的现场数据里称为移动性网络, 既包括运行状态的测试, 也包括运行状态测试的图像的测试模型。