In recent years novel architecture components for image classification have been developed, starting with attention and patches used in transformers. While prior works have analyzed the influence of some aspects of architecture components on the robustness to adversarial attacks, in particular for vision transformers, the understanding of the main factors is still limited. We compare several (non)-robust classifiers with different architectures and study their properties, including the effect of adversarial training on the interpretability of the learnt features and robustness to unseen threat models. An ablation from ResNet to ConvNeXt reveals key architectural changes leading to almost $10\%$ higher $\ell_\infty$-robustness.
翻译:近年来,开发了用于图像分类的新建筑构件,首先是在变压器中使用的注意力和补丁;先前的工程分析了建筑构件的某些方面对对抗性攻击的稳健性的影响,特别是对视觉变压器的影响,但对主要因素的了解仍然有限;我们比较了几个(非)强压分解器与不同结构并研究其特性,包括就所学特征的可解释性和对隐形威胁模型的稳健性进行对抗性培训的影响;从ResNet到ConneXt的合并揭示了重大的建筑变化,导致近10美元以上,即$\ell-infty$-robustness。