Inevitable domain and task discrepancies in real-world scenarios can impair the generalization performance of the pre-trained deep models for medical data. Therefore, we audaciously propose that we should build a general-purpose medical AI system that can be seamlessly adapted to downstream domains/tasks. Since the domain/task adaption procedures usually involve additional labeling work for the target data, designing a data-efficient adaption algorithm is desired to save the cost of transferring the learned knowledge. Our recent work found that vision-language models (VLMs) are efficient learners with extraordinary cross-domain ability. Therefore, in this work, we further explore the possibility of leveraging pre-trained VLMs as medical foundation models for building general-purpose medical AI, where we thoroughly investigate three machine-learning paradigms, i.e., domain/task-specialized learning, joint learning, and continual learning, for training the VLMs and evaluate their generalization performance on cross-domain and cross-task test sets. To alleviate the catastrophic forgetting during sequential training, we employ rehearsal learning and receive a sharp boost in terms of generalization capability. In a nutshell, our empirical evidence suggests that continual learning may be a practical and efficient learning paradigm for the medical foundation model. And we hope researchers can use our empirical evidence as basement to further explore the path toward medical foundation model.
翻译:由于域/任务适应程序通常涉及目标数据的额外标签工作,设计数据高效的适应算法是为了节省转让所学知识的成本。我们最近的工作发现,愿景语言模型(VLMS)是具有超常跨域能力的高效学习者。因此,在这项工作中,我们进一步探索利用预先培训的VLMs作为建立通用医疗AI的医疗基础模型的可能性,我们在该系统中彻底调查三种机器学习模式,即域/任务专门学习、联合学习和持续学习,以培训VLMs并评价其在跨域和跨任务测试中的总体性表现。为了减轻连续培训中的灾难性遗忘,我们利用彩排学习,并获得更大幅度的普及能力。在基因模型中,我们用经验证据来持续地探索医学基础。我们不断的实验性经验证据可以用来学习医学基础。</s>