Vision Transformers (ViT) are competing to replace Convolutional Neural Networks (CNN) for various computer vision tasks in medical imaging such as classification and segmentation. While the vulnerability of CNNs to adversarial attacks is a well-known problem, recent works have shown that ViTs are also susceptible to such attacks and suffer significant performance degradation under attack. The vulnerability of ViTs to carefully engineered adversarial samples raises serious concerns about their safety in clinical settings. In this paper, we propose a novel self-ensembling method to enhance the robustness of ViT in the presence of adversarial attacks. The proposed Self-Ensembling Vision Transformer (SEViT) leverages the fact that feature representations learned by initial blocks of a ViT are relatively unaffected by adversarial perturbations. Learning multiple classifiers based on these intermediate feature representations and combining these predictions with that of the final ViT classifier can provide robustness against adversarial attacks. Measuring the consistency between the various predictions can also help detect adversarial samples. Experiments on two modalities (chest X-ray and fundoscopy) demonstrate the efficacy of SEViT architecture to defend against various adversarial attacks in the gray-box (attacker has full knowledge of the target model, but not the defense mechanism) setting. Code: https://github.com/faresmalik/SEViT
翻译:视觉变异器(VIT)正在竞相取代进化神经网络(CNN),以取代医学成像(如分类和分割)中的各种计算机视觉任务。尽管CNN对对抗性攻击的脆弱性是一个众所周知的问题,但最近的工作表明,VET很容易受到这种攻击,并受到攻击,其性能严重退化。VIT对精心设计的对抗性样品的脆弱性引起人们对临床环境安全性的严重关切。在本文中,我们提议一种新型的自我聚合方法,以便在出现对抗性攻击时加强VT的稳健性。拟议的SEVT自我合成视觉变异器(SEVIT)利用了以下事实:VIT最初的区块所学的特征表现相对不受对抗性攻击的影响。学习基于这些中间特征的分类器,并将这些预测与最后VIT分类器的预测结合起来,可以提供抵御对抗性攻击性攻击的稳健性。测量各种预测的一致性也有助于检测对抗性攻击的样本。两种模式(胸前X光和花式)的实验展示了SEVIT的特征,显示SVIT初始/防御性攻击目标结构结构,用来防御性攻击的系统结构,用于防御性攻击。