Transfer learning from supervised ImageNet models has been frequently used in medical image analysis. Yet, no large-scale evaluation has been conducted to benchmark the efficacy of newly-developed pre-training techniques for medical image analysis, leaving several important questions unanswered. As the first step in this direction, we conduct a systematic study on the transferability of models pre-trained on iNat2021, the most recent large-scale fine-grained dataset, and 14 top self-supervised ImageNet models on 7 diverse medical tasks in comparison with the supervised ImageNet model. Furthermore, we present a practical approach to bridge the domain gap between natural and medical images by continually (pre-)training supervised ImageNet models on medical images. Our comprehensive evaluation yields new insights: (1) pre-trained models on fine-grained data yield distinctive local representations that are more suitable for medical segmentation tasks, (2) self-supervised ImageNet models learn holistic features more effectively than supervised ImageNet models, and (3) continual pre-training can bridge the domain gap between natural and medical images. We hope that this large-scale open evaluation of transfer learning can direct the future research of deep learning for medical imaging. As open science, all codes and pre-trained models are available on our GitHub page https://github.com/JLiangLab/BenchmarkTransferLearning.
翻译:在医疗图像分析中经常使用监督图像网模型的转移学习方法,然而,没有进行大规模评价,以衡量新开发的医学图像分析培训前技术的效率,留下几个重要问题没有得到答复。作为朝这个方向迈出的第一步,我们进行了系统研究,研究在 iNat2021 上预先培训的模型的可转让性,即最新的大规模微缩成份数据集,以及14个与监督的图像网模型相比在7种不同医疗任务上自我监督的图像网模型。此外,我们提出一种切实可行的办法,通过持续(预先)培训受监督的医学图像网络模型来弥合自然图像和医疗图像之间的领域差距。我们的全面评价产生了新的见解:(1) 预先培训的精细细重度数据模型产生独特的当地代表性,更适合医疗分解任务,(2) 自我监督的图像网模型比受监督的图像网模型更有效地学习整体特征,(3) 持续的培训前培训可以弥合自然和医疗图像之间的域差距。我们希望,这种大规模转移学习的公开评价能够指导我们今后对医学图像的深层学习模型进行研究。 开放的LFSO/Tregredustrals。