Increasing model, data and compute budget scale in the pre-training has been shown to strongly improve model generalization and transfer learning in vast line of work done in language modeling and natural image recognition. However, most studies on the positive effect of larger scale were done in scope of in-domain setting, with source and target data being in close proximity. To study effect of larger scale for both in-domain and out-of-domain setting when performing full and few-shot transfer, we combine here for the first time large, openly available medical X-Ray chest imaging datasets to reach a scale for medical imaging domain comparable to ImageNet-1k, routinely used for pre-training in natural image domain. We then conduct supervised pre-training, while varying network size and source data scale and domain, being either large natural (ImageNet-1k/21k) or large medical chest X-Ray datasets, and transfer pre-trained models to different natural or medical targets. We observe strong improvement due to larger pre-training scale for intra-domain natural-natural and medical-medical transfer. For inter-domain natural-medical transfer, we find improvements due to larger pre-training scale on larger X-Ray targets in full shot regime, while for smaller targets and for few-shot regime the improvement is not visible. Remarkably, large networks pre-trained on very large natural ImageNet-21k are as good or better than networks pre-trained on largest available medical X-Ray data when performing transfer to large X-Ray targets. We conclude that substantially increasing model and generic, medical domain-agnostic natural image source data scale in the pre-training can enable high quality out-of-domain transfer to medical domain specific targets, removing dependency on large medical domain-specific source data often not available in the practice.
翻译:培训前的模型、数据和预算规模的计算显示,在语言建模和自然图像识别方面的大量工作中,大大改进了模型的概括化和转移学习,但在语言建模和自然图像识别方面的大量工作中,大多数关于较大规模的积极影响的研究是在内部设置范围内进行的,而源和目标数据接近于此。为了研究在进行完整和少发传输时对内部和外部设置的较大规模的影响,我们首次在这里合并了大规模、公开提供的医学X光胸前成像数据集集,以达到与图像Net-1k相类似的医学成像域域域域域域的升级规模。我们随后在自然图像域域域域域域前进行监管前的大型规模和源数据规模和域域,无论是大型自然(ImageNet-1k/21k)或大型的X光胸前型数据集配置,还是各种经过训练的模型转换到不同的自然或医疗目标。我们观察到,由于较大规模的培训前大型内部和医疗机域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域域内现有规模的升级的升级的升级,我们无法进行更甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚,因此,因此,因此,在大规模甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚甚