Self supervised contrastive learning based pretraining allows development of robust and generalized deep learning models with small, labeled datasets, reducing the burden of label generation. This paper aims to evaluate the effect of CL based pretraining on the performance of referrable vs non referrable diabetic retinopathy (DR) classification. We have developed a CL based framework with neural style transfer (NST) augmentation to produce models with better representations and initializations for the detection of DR in color fundus images. We compare our CL pretrained model performance with two state of the art baseline models pretrained with Imagenet weights. We further investigate the model performance with reduced labeled training data (down to 10 percent) to test the robustness of the model when trained with small, labeled datasets. The model is trained and validated on the EyePACS dataset and tested independently on clinical data from the University of Illinois, Chicago (UIC). Compared to baseline models, our CL pretrained FundusNet model had higher AUC (CI) values (0.91 (0.898 to 0.930) vs 0.80 (0.783 to 0.820) and 0.83 (0.801 to 0.853) on UIC data). At 10 percent labeled training data, the FundusNet AUC was 0.81 (0.78 to 0.84) vs 0.58 (0.56 to 0.64) and 0.63 (0.60 to 0.66) in baseline models, when tested on the UIC dataset. CL based pretraining with NST significantly improves DL classification performance, helps the model generalize well (transferable from EyePACS to UIC data), and allows training with small, annotated datasets, therefore reducing ground truth annotation burden of the clinicians.
翻译:以自我监督的对比学习为基础的训练前,可以开发强有力和通用的深层次学习模型,并配以有标签的小型数据集,减少标签生成的负担。本文旨在评估基于CL的训练前,对可参考性和非可参考性糖尿病视像病(DR)分类的性能的影响。我们开发了基于CL的框架,配有神经风格传输(NST)增强,制成具有更好表现和初始化的模型,用于检测彩色基金图像中的DR。我们比较了CL预先培训的模型性能与两个先进的艺术基线模型的性能状态进行了对比。我们进一步调查模型的性能,用显著的标签培训数据(降至10% ) 来测试该模型的可靠性。我们用EepPACS数据集进行了培训和独立测试,并用伊利诺伊大学的临床数据进行测试。与基线模型相比,我们的CLLFNet模型的性能测试前(CI)值从0.91 (0.898)到0.930 (0.80) 和0.80(0.783 ST 到0.8) 基本数据,因此,用O.ILA.01到0.86 和0.688 数据库数据。