While deep neural networks have shown impressive performance in many tasks, they are fragile to carefully designed adversarial attacks. We propose a novel adversarial training-based model by Attention Guided Knowledge Distillation and Bi-directional Metric Learning (AGKD-BML). The attention knowledge is obtained from a weight-fixed model trained on a clean dataset, referred to as a teacher model, and transferred to a model that is under training on adversarial examples (AEs), referred to as a student model. In this way, the student model is able to focus on the correct region, as well as correcting the intermediate features corrupted by AEs to eventually improve the model accuracy. Moreover, to efficiently regularize the representation in feature space, we propose a bidirectional metric learning. Specifically, given a clean image, it is first attacked to its most confusing class to get the forward AE. A clean image in the most confusing class is then randomly picked and attacked back to the original class to get the backward AE. A triplet loss is then used to shorten the representation distance between original image and its AE, while enlarge that between the forward and backward AEs. We conduct extensive adversarial robustness experiments on two widely used datasets with different attacks. Our proposed AGKD-BML model consistently outperforms the state-of-the-art approaches. The code of AGKD-BML will be available at: https://github.com/hongw579/AGKD-BML.
翻译:虽然深神经网络在许多任务中表现出了令人印象深刻的绩效,但它们对于精心设计的对抗性攻击来说是脆弱的。我们提出一个新的对抗性培训模式,由“注意引导知识蒸馏”和双向计量学习(AGKD-BML)组成。关注知识来自在清洁数据集方面受过培训的重力固定模型,被称为教师模式,并被转移到正在接受关于对抗性实例(AEs)的培训的模型,被称为学生模式。这样,学生模式能够侧重于正确的区域,并纠正由AEs腐蚀的中间特征,以便最终提高模型的准确性。此外,为了有效地规范地貌空间的表述,我们建议采用双向计量学习。具体地说,如果有一个干净的图像,它首先被攻击到最令人困惑的类别,然后被随机挑选,然后又被攻击回到最初的类别,以获得落后的AE。三重损失被用来缩短原始图像和AEE的中间特征。我们广泛缩小了原始图像和AE之间的代表距离,同时扩大前向和后向式ABK的实验。我们使用的ABK-AF-A-A-A-A-A-C-C-A-A-A-C-A-A-A-A-A-A-A-A-A-A-C-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-