Computer aided diagnosis (CAD) has gained an increased amount of attention in the general research community over the last years as an example of a typical limited data application - with experiments on labeled 100k-200k datasets. Although these datasets are still small compared to natural image datasets like ImageNet1k, ImageNet21k and JFT, they are large for annotated medical datasets, where 1k-10k labeled samples are much more common. There is no baseline on which methods to build on in the low data regime. In this work we bridge this gap by providing an extensive study on medical image classification with limited annotations (5k). We present a study of modern architectures applied to a fixed low data regime of 5000 images on the CheXpert dataset. Conclusively we find that models pretrained on ImageNet21k achieve a higher AUC and larger models require less training steps. All models are quite well calibrated even though we only fine-tuned on 5000 training samples. All 'modern' architectures have higher AUC than ResNet50. Regularization of Big Transfer Models with MixUp or Mean Teacher improves calibration, MixUp also improves accuracy. Vision Transformer achieve comparable or on par results to Big Transfer Models.
翻译:过去几年来,计算机辅助诊断(CAD)作为典型有限数据应用的典型有限数据应用实例,在100k-200k数据集上进行了实验。尽管与图像Net1k、图像Net21k和JFT等自然图像数据集相比,这些数据集仍然很小,但它们对于附加说明的医疗数据集而言却很大,其中1k-10k标签标签样本更为常见。在低数据系统中,没有根据哪些方法进行改进的基线。在这项工作中,我们通过提供带有有限说明的医学图像分类的广泛研究(5k)来弥补这一差距。我们介绍了对适用于CheXpert数据集5000个图像的固定低数据系统的现代结构进行的一项研究。我们一致地发现,在图像Net21k上预先培训的模型达到更高AUC和更大的模型需要较少的培训步骤。所有模型都非常精确,尽管我们只对5000个培训样本进行微调。所有“现代”结构都比ResNet50高。用MixUP或BIGMIX模型改进了大或可比较的图像校准。