We show that label noise exists in adversarial training. Such label noise is due to the mismatch between the true label distribution of adversarial examples and the label inherited from clean examples - the true label distribution is distorted by the adversarial perturbation, but is neglected by the common practice that inherits labels from clean examples. Recognizing label noise sheds insights on the prevalence of robust overfitting in adversarial training, and explains its intriguing dependence on perturbation radius and data quality. Also, our label noise perspective aligns well with our observations of the epoch-wise double descent in adversarial training. Guided by our analyses, we proposed a method to automatically calibrate the label to address the label noise and robust overfitting. Our method achieves consistent performance improvements across various models and datasets without introducing new hyper-parameters or additional tuning.
翻译:我们发现,在对抗性训练中存在标签噪音,这种标签噪音是由于对抗性例子的真正标签分布与从干净例子中继承的标签之间的不匹配,真正的标签分布被对抗性干扰扭曲,但被继承标签的常见做法所忽视,从干净的例子中继承标签。认识到标签噪音使人们对激烈过度装配在对抗性训练中的普遍程度有了深刻的认识,并解释了它对扰动半径和数据质量的令人感兴趣的依赖。此外,我们的标签噪音观点与我们在对抗性训练中粗巧的双向下降的观察非常一致。我们的分析指出,我们建议了一种方法来自动校准标签,以解决标签噪音和稳健的过度装配。我们的方法在不引入新的超参数或额外调制的情况下,在各种模型和数据集中实现了一致的性能改进。