Beyond achieving high performance across many vision tasks, multimodal models are expected to be robust to single-source faults due to the availability of redundant information between modalities. In this paper, we investigate the robustness of multimodal neural networks against worst-case (i.e., adversarial) perturbations on a single modality. We first show that standard multimodal fusion models are vulnerable to single-source adversaries: an attack on any single modality can overcome the correct information from multiple unperturbed modalities and cause the model to fail. This surprising vulnerability holds across diverse multimodal tasks and necessitates a solution. Motivated by this finding, we propose an adversarially robust fusion strategy that trains the model to compare information coming from all the input sources, detect inconsistencies in the perturbed modality compared to the other modalities, and only allow information from the unperturbed modalities to pass through. Our approach significantly improves on state-of-the-art methods in single-source robustness, achieving gains of 7.8-25.2% on action recognition, 19.7-48.2% on object detection, and 1.6-6.7% on sentiment analysis, without degrading performance on unperturbed (i.e., clean) data.
翻译:除了在许多愿景任务中实现高绩效之外,多式联运模型预计对单一来源缺陷的单一来源缺陷具有强力作用,因为各种模式之间存在多余的信息。在本文件中,我们调查了多式联运神经网络对于一种单一模式最坏(即对抗性)干扰的强力性。我们首先表明,标准的多式联运聚合模式很容易被单一来源对手所利用:对任何单一模式的攻击都能够克服多个未受干扰模式的正确信息,并导致该模式失败。这种令人惊讶的脆弱性贯穿于多种模式的任务中,需要找到解决办法。我们基于这一发现,建议了一种对抗性强健的聚合战略,以对来自所有投入来源的信息进行比较,发现受干扰模式与其他模式的不一致之处,并只允许来自未受扰动模式的信息通过。我们的方法大大改进了单一来源稳健的先进方法,实现了7.8%至25.2%的行动识别收益,19.7%至48.2%的物体检测成果和1.6至6.7%的情绪分析结果,而没有降低未受扰动数据的业绩(i)。