Fundus photography is the primary method for retinal imaging and essential for diabetic retinopathy prevention. Automated segmentation of fundus photographs would improve the quality, capacity, and cost-effectiveness of eye care screening programs. However, current segmentation methods are not robust towards the diversity in imaging conditions and pathologies typical for real-world clinical applications. To overcome these limitations, we utilized contrastive self-supervised learning to exploit the large variety of unlabeled fundus images in the publicly available EyePACS dataset. We pre-trained an encoder of a U-Net, which we later fine-tuned on several retinal vessel and lesion segmentation datasets. We demonstrate for the first time that by using contrastive self-supervised learning, the pre-trained network can recognize blood vessels, optic disc, fovea, and various lesions without being provided any labels. Furthermore, when fine-tuned on a downstream blood vessel segmentation task, such pre-trained networks achieve state-of-the-art performance on images from different datasets. Additionally, the pre-training also leads to shorter training times and an improved few-shot performance on both blood vessel and lesion segmentation tasks. Altogether, our results showcase the benefits of contrastive self-supervised pre-training which can play a crucial role in real-world clinical applications requiring robust models able to adapt to new devices with only a few annotated samples.
翻译:基金摄影是视网膜成像的首要方法,也是糖尿病视网膜病原体预防必不可少的基本方法。基金照片的自动分解将提高眼睛护理筛查方案的质量、能力和成本效益。然而,目前的分解方法对于真实世界临床应用中典型的成像条件和病理的多样化并不十分健全。为了克服这些限制,我们利用对比式自我监督学习来利用公众可公开查阅的EyePACS数据集中大量未贴标签的基金图像。我们预先训练了一个U-Net的样本编码器,我们后来对一些视网膜容器和偏差分解数据集进行了微调。我们第一次展示了通过对比式自我监督学习,预先训练的网络可以识别血管容器、光碟、软盘和各种损伤,而没有提供任何标签。此外,在对下游血管分解模型进行微调时,这种预先训练的网络只能对不同数据集的图像进行最先进的性能表现。此外,在培训前的临床应用中,也只能通过短期的自我监督性能和自我训练,从而可以提高一个更精确的容器的自我训练阶段。