Deep neural networks (DNNs) are under threat from adversarial examples. Adversarial detection is a fundamental work for robust DNNs-based service, which distinguishes adversarial images from benign images. Image transformation is one of the most effective approaches to detect adversarial examples. During the last few years, a variety of image transformations have been studied and discussed to design reliable adversarial detectors. In this paper, we systematically review the recent progress on adversarial detection via image transformations with a novel taxonomy. Then we conduct an extensive set of experiments to test the detection performance of image transformations towards the state-of-the-art adversarial attacks. Furthermore, we reveal that the single transformation is not capable of detecting robust adversarial examples, and propose an improved approach by combining multiple image transformations. The results show that the joint approach achieves significant improvement in detection accuracy and recall. We suggest that the joint detector is a more effective tool to detect adversarial examples.
翻译:深神经网络(DNN)受到对抗性实例的威胁。 反向检测是强力 DNNs 基础服务的基本工作,它区分了对抗性图像和良性图像。 图像转换是发现对抗性实例的最有效方法之一。 在过去几年里,对各种图像转换进行了研究和讨论,以设计可靠的对抗性探测器。 在本文中,我们系统地审查最近通过新分类法的图像转换在对抗性检测方面取得的进展。 然后,我们进行了一系列广泛的实验,以测试向最先进的对抗性攻击转变图像的检测性能。 此外,我们发现单项转换无法发现有力的对抗性实例,并通过合并多个图像转换提出改进的方法。结果显示,联合方法在探测准确性方面有显著改进,并忆及。 我们建议,联合检测器是发现对抗性实例的更有效工具。