The introduction of robust optimisation has pushed the state-of-the-art in defending against adversarial attacks. However, the behaviour of such optimisation has not been studied in the light of a fundamentally different class of attacks called backdoors. In this paper, we demonstrate that adversarially robust models are susceptible to backdoor attacks. Subsequently, we observe that backdoors are reflected in the feature representation of such models. Then, this observation is leveraged to detect backdoor-infected models via a detection technique called AEGIS. Specifically, AEGIS uses feature clustering to effectively detect backdoor-infected robust Deep Neural Networks (DNNs). In our evaluation of several visible and hidden backdoor triggers on major classification tasks using CIFAR-10, MNIST and FMNIST datasets, AEGIS effectively detects robust DNNs infected with backdoors. AEGIS detects a backdoor-infected model with 91.6% accuracy, without any false positives. Furthermore, AEGIS detects the targeted class in the backdoor-infected model with a reasonably low (11.1%) false positive rate. Our investigation reveals that salient features of adversarially robust DNNs break the stealthy nature of backdoor attacks.
翻译:强力优化的引入推动了防范对抗性攻击的最先进技术的采用。然而,这种优化的行为还没有根据一个完全不同的、称为后门的攻击类别进行研究。在本文中,我们证明敌对性强的模型很容易受到后门攻击。随后,我们观察到后门在这种模型的特征中有所反映。然后,利用这一观察,通过一种称为AEGIS的探测技术,检测后门感染的模式。具体地说,AEGIS利用特征集群,有效检测后门感染的强大深神经网络(DNN)。在我们对使用CIFAR-10、MNIST和FMNIST数据集进行的主要分类任务的若干可见和隐藏的后门触发器的评估中,AEGIS有效地检测出有后门感染的强势DNN。AGIS检测出一种准确度为91.6%的后门感染模式,而没有任何虚假的阳性。此外,AEGIS检测后门感染模式中的目标类别,以相当低的(11.1%)错误的反向积极率。我们的调查揭示了隐性攻击的性质。