Adversarial training aims to reduce the problematic susceptibility of modern neural networks to small data perturbations. Surprisingly, overfitting is a major concern in adversarial training of neural networks despite being mostly absent in standard training. We provide here theoretical evidence for this peculiar ``robust overfitting'' phenomenon. Subsequently, we advance a novel loss function which we show both theoretically as well as empirically to enjoy a certified level of robustness against data evasion and poisoning attacks while ensuring guaranteed generalization. We indicate through careful numerical experiments that our resulting holistic robust (HR) training procedure yields SOTA performance in terms of adversarial error loss. Finally, we indicate that HR training can be interpreted as a direct extension of adversarial training and comes with a negligible additional computational burden.
翻译:反向培训旨在降低现代神经网络对小型数据扰动的棘手易感性; 令人惊讶的是,尽管大多没有参加标准培训,但过度装配是神经网络对抗性培训的一个主要问题; 我们在此为这种奇特的“罗布斯特超称”现象提供理论证据; 随后, 我们推出一种新的损失功能, 我们在理论上和经验上都表明,在防止数据规避和中毒攻击的同时,在确保普遍化的同时,享有经认证的稳健程度; 我们通过仔细的数字实验表明,我们由此形成的整体强健(HR)培训程序在对抗性错误损失方面产生了SOTA的性能。 最后,我们指出,人力资源培训可以被解释为对抗性培训的直接延伸,并且带有微不足道的额外计算负担。</s>