Medical Visual Question Answering (Medical-VQA) aims to answer clinical questions regarding radiology images, assisting doctors with decision-making options. Nevertheless, current Medical-VQA models learn cross-modal representations through residing vision and texture encoders in dual separate spaces, which lead to indirect semantic alignment. In this paper, we propose UnICLAM, a Unified and Interpretable Medical-VQA model through Contrastive Representation Learning with Adversarial Masking. Specifically, to learn an aligned image-text representation, we first establish a unified dual-stream pre-training structure with the gradually soft-parameter sharing strategy. Technically, the proposed strategy learns a constraint for the vision and texture encoders to be close in a same space, which is gradually loosened as the higher number of layers. Moreover, for grasping the semantic representation, we extend the unified Adversarial Masking data augmentation strategy to the contrastive representation learning of vision and text in a unified manner, alleviating the meaningless of the commonly used random mask. Concretely, while the encoder training minimizes the distance between the original feature and the masking feature, the adversarial masking model keeps adversarial learning to conversely maximize the distance. Furthermore, we also intuitively take a further exploration of the unified adversarial masking strategy, which improves the potential ante-hoc interpretability with remarkable performance and efficiency. Experimental results on VQA-RAD and SLAKE public benchmarks demonstrate that UnICLAM outperforms the existing 11 state-of-the-art Medical-VQA models. More importantly, we make an additional discussion about the performance of UnICLAM in diagnosing heart failure, verifying that UnICLAM exhibits superior few-shot adaption performance in practical disease diagnosis.
翻译:医学视觉问答(Medical-VQA)旨在解答有关放射图像的临床问题,协助医生做出决策选择;然而,目前的医学VQA模型通过在双独立空间的内置视觉和纹理编码器学习跨模式表现,这导致间接的语义一致。在本文中,我们建议UniICLAM,一个通过反向遮罩进行对比演示学习的统一和解释的医学-VQA模型。具体地说,为了学习一致的图像文本代表制,我们首先建立一个统一的双流培训前结构,同时采用逐渐软参数共享战略。技术上,拟议的战略通过在双独立空间学习双向图像和纹理编码,学习跨双向的视觉和纹理描述,同时,对视觉和纹理进行更深层次的演示。