In this paper, we ask whether Vision Transformers (ViTs) can serve as an underlying architecture for improving the adversarial robustness of machine learning models against evasion attacks. While earlier works have focused on improving Convolutional Neural Networks, we show that also ViTs are highly suitable for adversarial training to achieve competitive performance. We achieve this objective using a custom adversarial training recipe, discovered using rigorous ablation studies on a subset of the ImageNet dataset. The canonical training recipe for ViTs recommends strong data augmentation, in part to compensate for the lack of vision inductive bias of attention modules, when compared to convolutions. We show that this recipe achieves suboptimal performance when used for adversarial training. In contrast, we find that omitting all heavy data augmentation, and adding some additional bag-of-tricks ($\varepsilon$-warmup and larger weight decay), significantly boosts the performance of robust ViTs. We show that our recipe generalizes to different classes of ViT architectures and large-scale models on full ImageNet-1k. Additionally, investigating the reasons for the robustness of our models, we show that it is easier to generate strong attacks during training when using our recipe and that this leads to better robustness at test time. Finally, we further study one consequence of adversarial training by proposing a way to quantify the semantic nature of adversarial perturbations and highlight its correlation with the robustness of the model. Overall, we recommend that the community should avoid translating the canonical training recipes in ViTs to robust training and rethink common training choices in the context of adversarial training.
翻译:在本文中,我们询问View Trangers(View Trangers)能否成为改善机器学习模式对抗规避攻击的对抗性强势基础架构。 虽然先前的工作重点是改善革命神经网络, 但我们也显示ViT非常适合进行对抗性培训, 以取得竞争性的绩效。 我们使用一个定制的对抗性培训食谱, 在图像网络数据集的子集上采用严格的反动研究发现, 严格反动性培训食谱, 从而大大提升了强势ViT的绩效。 维Ts的典型培训食谱建议了强大的数据扩充, 部分是为了弥补在与演进相比, 关注模块缺乏直观性倾向。 我们表明,在使用对抗性培训时, 也取得了不理想性的表现。 相比之下, 研究更稳健健性培训的原因, 最终在测试模型时, 将更稳健性的培训结果转化为更稳健性的培训。 我们的良性培训方式, 最终将更能转化为更稳健性的培训。