Adversarial training has been demonstrated to be one of the most effective remedies for defending adversarial examples, yet it often suffers from the huge robustness generalization gap on unseen testing adversaries, deemed as the \emph{adversarially robust generalization problem}. Despite the preliminary understandings devoted on adversarially robust generalization, little is known from the architectural perspective. Thus, this paper tries to bridge the gap by systematically examining the most representative architectures (e.g., Vision Transformers and CNNs). In particular, we first comprehensively evaluated \emph{20} adversarially trained architectures on ImageNette and CIFAR-10 datasets towards several adversaries (multiple $\ell_p$-norm adversarial attacks), and found that Vision Transformers (e.g., PVT, CoAtNet) often yield better adversarially robust generalization. To further understand what architectural ingredients favor adversarially robust generalization, we delve into several key building blocks and revealed the fact via the lens of Rademacher complexity that the higher weight sparsity contributes significantly towards the better adversarially robust generalization of Vision Transformers, which can be often achieved by attention layers. Our extensive studies discovered the close relationship between architectural design and adversarially robust generalization, and instantiated several important insights. We hope our findings could help to better understand the mechanism towards designing robust deep learning architectures.
翻译:实践证明,Aversarial Adversarial 培训是捍卫对抗性实例的最有效补救措施之一,然而,它往往会因隐蔽的测试对手(被视作对抗性激烈的对抗性攻击)的强大一般化差距而受害。 尽管初步的谅解致力于对抗性强的概括化,但从建筑角度看却鲜为人知。因此,本文件试图通过系统检查最具代表性的结构(例如,愿景变换器和CNN)来弥合差距。特别是,我们首先全面评估了在图像网和CIFAR-10上经过对抗性训练的架构,这些架构对一些对手(被视作mulple $@ell_p$-norm对抗性攻击)有着巨大的强势一般化差距。 该文件还发现,愿景变换器(例如,PVT, CoAtNet)往往产生更好的对抗性强势概括化认识。 为了进一步理解哪些建筑要素有利于对抗性强势的概括化,我们分解了几个关键构件,通过Redemacher 复杂度的视角揭示了事实,即更高的重度关注度有助于更好地实现对立性更稳健的深刻的愿景变型结构设计。