Deep learning (DL) has shown great success in many human-related tasks, which has led to its adoption in many computer vision based applications, such as security surveillance systems, autonomous vehicles and healthcare. Such safety-critical applications have to draw their path to success deployment once they have the capability to overcome safety-critical challenges. Among these challenges are the defense against or/and the detection of the adversarial examples (AEs). Adversaries can carefully craft small, often imperceptible, noise called perturbations to be added to the clean image to generate the AE. The aim of AE is to fool the DL model which makes it a potential risk for DL applications. Many test-time evasion attacks and countermeasures,i.e., defense or detection methods, are proposed in the literature. Moreover, few reviews and surveys were published and theoretically showed the taxonomy of the threats and the countermeasure methods with little focus in AE detection methods. In this paper, we focus on image classification task and attempt to provide a survey for detection methods of test-time evasion attacks on neural network classifiers. A detailed discussion for such methods is provided with experimental results for eight state-of-the-art detectors under different scenarios on four datasets. We also provide potential challenges and future perspectives for this research direction.
翻译:深入学习(DL)在许多与人类有关的任务中表现出了巨大的成功,这导致它在许多基于计算机的视觉应用中被采纳,例如安全监视系统、自主车辆和保健等。这类安全关键应用一旦有能力克服安全关键挑战,就必须走上成功部署的道路。这些挑战包括防患于未然的防御或/和辨别对抗实例(AEs),对立面可以谨慎地设计小的、往往不易察觉的噪音,称为扰动,以生成AE。AE的目的是欺骗DL模型,使DL成为DL应用的潜在风险。文献中提出了许多测试性规避攻击和反措施,即防御或探测方法。此外,很少公布和理论上展示威胁的分类和反制方法,而AE探测方法很少受到重视。在本文中,我们侧重于图像分类任务,并试图为探测对神经网络分类人员进行测试性规避攻击的方法提供调查。我们用四种方法进行详细讨论,以实验性的方式对八种状态趋势进行定位。