Recent advances in vision-language pre-training have demonstrated astounding performances in diverse vision-language tasks, shedding a light on the long-standing problems of a comprehensive understanding of both visual and textual concepts in artificial intelligence research. However, there has been limited success in the application of vision-language pre-training in the medical domain, as the current vision-language models and learning strategies for photographic images and captions are not optimal to process the medical data which are usually insufficient in the amount and the diversity, which impedes successful learning of joint vision-language concepts. In this study, we introduce MAX-VL, a model tailored for efficient vision-language pre-training in the medical domain. We experimentally demonstrated that the pre-trained MAX-VL model outperforms the current state-of-the-art vision language models in various vision-language tasks. We also suggested the clinical utility for the diagnosis of newly emerging diseases and human error detection as well as showed the widespread applicability of the model in different domain data.
翻译:最近在视力语言培训前取得的进展表明,在各种视力语言任务中取得了惊人的成绩,揭示了在全面理解人工智能研究中的视觉和文字概念方面长期存在的问题,然而,在医疗领域应用视觉语言培训前的成绩有限,因为目前的视觉语言照片和字幕模型和学习战略对于处理医学数据来说并不理想,而医学数据通常数量不足,多样性也妨碍成功学习联合视力语言概念。在本研究中,我们引入了MAX-VL,这是专门为医学领域高效视觉语言培训前培训而设计的模型。我们实验性地证明,预先培训的MAX-VL模型在各种视觉语言任务中超越了目前最先进的视觉语言模型。我们还提出了诊断新出现疾病和人类误差的临床效用,并表明该模型在不同的领域数据中广泛适用。