The vulnerability of deep neural networks (DNNs) to adversarial examples has attracted great attention in the machine learning community. The problem is related to non-flatness and non-smoothness of normally obtained loss landscapes. Training augmented with adversarial examples (a.k.a., adversarial training) is considered as an effective remedy. In this paper, we highlight that some collaborative examples, nearly perceptually indistinguishable from both adversarial and benign examples yet show extremely lower prediction loss, can be utilized to enhance adversarial training. A novel method is therefore proposed to achieve new state-of-the-arts in adversarial robustness. Code: https://github.com/qizhangli/ST-AT.
翻译:深层神经网络(DNN)易受对抗性实例的影响,这在机器学习界引起了极大关注,问题与通常获得的损失场景的无负缩缩和不松动有关,培训加上对抗性实例(a.k.a.a.,对抗性培训)被视为一种有效的补救办法,在本文件中,我们强调,一些合作实例,几乎无法与对抗性和良性实例区分开来,但预测损失却极低,可以用来加强对抗性培训,因此提出了在对抗性强力方面实现新状态的新办法,守则:https://github.com/qizhangli/ST-AT。