Today, a large number of glaucoma cases remain undetected, resulting in irreversible blindness. In a quest for cost-effective screening, deep learning-based methods are being evaluated to detect glaucoma from color fundus images. Although unprecedented sensitivity and specificity values are reported, recent glaucoma detection deep learning models lack in decision transparency. Here, we propose a methodology that advances explainable deep learning in the field of glaucoma detection and vertical cup-disc ratio (VCDR), an important risk factor. We trained and evaluated a total of 64 deep learning models using fundus images that undergo a certain cropping policy. We defined the circular crop radius as a percentage of image size, centered on the optic nerve head (ONH), with an equidistant spaced range from 10%-60% (ONH crop policy). The inverse of the cropping mask was also applied to quantify the performance of models trained on ONH information exclusively (periphery crop policy). The performance of the models evaluated on original images resulted in an area under the curve (AUC) of 0.94 [95% CI: 0.92-0.96] for glaucoma detection, and a coefficient of determination (R^2) equal to 77% [95% CI: 0.77-0.79] for VCDR estimation. Models that were trained on images with absence of the ONH are still able to obtain significant performance (0.88 [95% CI: 0.85-0.90] AUC for glaucoma detection and 37% [95% CI: 0.35-0.40] R^2 score for VCDR estimation in the most extreme setup of 60% ONH crop). We validated our glaucoma detection models on a recent public data set (REFUGE) that contains images captured with a different camera, still achieving an AUC of 0.80 [95% CI: 0.76-0.84] when ONH crop policy of 60% image size was applied. Our findings provide the first irrefutable evidence that deep learning can detect glaucoma from fundus image regions outside the ONH.
翻译:今天,大量青光眼病例仍未被发现,导致不可逆转的失明。在寻求具有成本效益的筛选过程中,正在对深层次的学习方法进行评估,以便从彩色基金图像中检测青光眼。尽管报告了前所未有的敏感度和特殊性值,但最近青光眼检测深度学习模型缺乏决策透明度。在这里,我们提出了一个方法,在青光眼检测和垂直杯状比(VCDR)领域推进可以解释的深层次学习,这是一个重要风险因素。我们培训和评估了总共64个深层次学习模型,使用了经过某种裁剪切政策的基金图像。我们把圆形作物检测半径定义为图像大小的一个百分比,以光学神经模型(ONH)为中心,最近青光眼发现模型为10-60%(ONH作物政策)。我们用作物口罩来量化仅用OH信息培训的模型的性能[OREULOFI政策 。我们用原始图像的性能评估结果在0.94 [95 CLOR-05) 底线下区域,用0.95 直径数据检测结果为ODR:0.95 内测算结果,用直径为ODR.96。我们测算值的数值为OMI值的数值为OM值的数值为0.9%。我们测算的数值的数值为OM值的数值的数值的数值为0.0.95-05,用于等的数值的数值为OM值。