Transfer learning has gained attention in medical image analysis due to limited annotated 3D medical datasets for training data-driven deep learning models in the real world. Existing 3D-based methods have transferred the pre-trained models to downstream tasks, which achieved promising results with only a small number of training samples. However, they demand a massive amount of parameters to train the model for 3D medical imaging. In this work, we propose a novel transfer learning framework, called Medical Transformer, that effectively models 3D volumetric images in the form of a sequence of 2D image slices. To make a high-level representation in 3D-form empowering spatial relations better, we take a multi-view approach that leverages plenty of information from the three planes of 3D volume, while providing parameter-efficient training. For building a source model generally applicable to various tasks, we pre-train the model in a self-supervised learning manner for masked encoding vector prediction as a proxy task, using a large-scale normal, healthy brain magnetic resonance imaging (MRI) dataset. Our pre-trained model is evaluated on three downstream tasks: (i) brain disease diagnosis, (ii) brain age prediction, and (iii) brain tumor segmentation, which are actively studied in brain MRI research. The experimental results show that our Medical Transformer outperforms the state-of-the-art transfer learning methods, efficiently reducing the number of parameters up to about 92% for classification and
翻译:由于在现实世界中培训以数据驱动的深层次学习模型的附加说明的3D医学数据集有限,医学图像分析中转移学习受到关注。现有的3D方法已经将预先培训的模式转移到下游任务中,这些任务只用少量培训样本取得了有希望的成果。然而,它们需要大量参数来培训3D医学成像模型。在这项工作中,我们提出了一个名为医疗变异器的新转移学习框架,以2D图像切片序列的形式,有效地模型3D体积图像。为了在3D形态增强空间关系中提高高级别代表性,我们采取了多视角方法,利用3D体积三平层的大量信息,同时提供节能培训。为了建立一个普遍适用于各种任务的源码模型,我们先用自我监督的学习模式,将遮蔽的编码矢量预测作为一种代理任务,使用大规模正常、健康的脑磁共振荡成像(MRI)数据集。我们预先培训的模型正在评估三项下游任务:(i) 大脑疾病诊断,(ii) 大脑年龄分析,并(ii) 测试大脑诊断,大脑诊断,(ii) 大脑年龄分析,(ralformastration) 进行积极的大脑诊断,大脑分析。