Supervised deep learning algorithms hold great potential to automate screening, monitoring and grading of medical images. However, training performant models has typically required vast quantities of labelled data, which is scarcely available in the medical domain. Self-supervised contrastive frameworks relax this dependency by first learning from unlabelled images. In this work we show that pretraining with two contrastive methods, SimCLR and BYOL, improves the utility of deep learning with regard to the clinical assessment of age-related macular degeneration (AMD). In experiments using two large clinical datasets containing 170,427 optical coherence tomography (OCT) images of 7,912 patients, we evaluate benefits attributed to pretraining across seven downstream tasks ranging from AMD stage and type classification to prediction of functional endpoints to segmentation of retinal layers, finding performance significantly increased in six out of seven tasks with fewer labels. However, standard contrastive frameworks have two known weaknesses that are detrimental to pretraining in the medical domain. Several of the image transformations used to create positive contrastive pairs are not applicable to greyscale medical scans. Furthermore, medical images often depict the same anatomical region and disease severity, resulting in numerous misleading negative pairs. To address these issues we develop a novel metadata-enhanced approach that exploits the rich set of inherently available patient information. To this end we employ records for patient identity, eye position (i.e. left or right) and time series data to indicate the typically unknowable set of inter-image contrastive relationships. By leveraging this often neglected information our metadata-enhanced contrastive pretraining leads to further benefits and outperforms conventional contrastive methods in five out of seven downstream tasks.
翻译:受监督的深层学习算法具有使医学图像自动筛选、监测和定级的巨大潜力。然而,培训表现模型通常需要大量贴标签的数据,而医疗领域却很少能获得这些数据。自我监督的对比框架通过首先从未贴标签的图像中学习,放松了这种依赖性。在这项工作中,我们通过两种对比方法,即SimCLR和BYOL, 提高了在临床评估与年龄有关的肌肉退化(AMD)方面深层学习的效用。在使用含有7,912病人的170,427个光学一致性图像(OCT)的两个大型临床数据集的实验中,我们通常需要评估从AMD阶段和类型分类到功能端端点的预测等七个下游任务的预培训所带来的好处。我们发现,在7项任务中,SimCLR和BYOL的绩效明显提高。然而,标准对比框架有两种已知的弱点,不利于医学领域前期培训。一些用于创建正对比配的图像转换关系,往往不适用于非灰度医学位置的右侧图像扫描。此外,我们往往要用这些老式的图像来分析系统记录。