Robustness of machine learning models to various adversarial and non-adversarial corruptions continues to be of interest. In this paper, we introduce the notion of the boundary thickness of a classifier, and we describe its connection with and usefulness for model robustness. Thick decision boundaries lead to improved performance, while thin decision boundaries lead to overfitting (e.g., measured by the robust generalization gap between training and testing) and lower robustness. We show that a thicker boundary helps improve robustness against adversarial examples (e.g., improving the robust test accuracy of adversarial training) as well as so-called out-of-distribution (OOD) transforms, and we show that many commonly-used regularization and data augmentation procedures can increase boundary thickness. On the theoretical side, we establish that maximizing boundary thickness during training is akin to the so-called mixup training. Using these observations, we show that noise-augmentation on mixup training further increases boundary thickness, thereby combating vulnerability to various forms of adversarial attacks and OOD transforms. We can also show that the performance improvement in several lines of recent work happens in conjunction with a thicker boundary.
翻译:机器学习模式对各种对抗性和非对抗性腐败的坚固性仍然令人感兴趣。 在本文中,我们引入了分类者的边界厚度概念,并描述了其与模型稳健度的联系和实用性。厚度决定边界导致性能的改善,而薄度决定边界则导致超标(例如,以严格的培训和测试之间的普遍化差距来衡量)和低强度。我们表明,较厚的边界有助于增强抵御对抗性例子(例如,提高对抗性训练的稳健测试准确性)以及所谓的分配外变(OOOD)的稳健性,我们还表明,许多常用的正规化和数据增强程序可以增加边界厚度。在理论方面,我们确定,培训期间最大程度的边界厚度与所谓的混合培训相似。我们利用这些观察,表明混合培训的噪音放大进一步增加了边界厚度,从而消除了对各种形式的对抗性攻击和OOD变换的脆弱性。我们还可以表明,最近工作的若干行的绩效改进与厚度边界同时发生。