Real-world non-mydriatic retinal fundus photography is prone to artifacts, imperfections and low-quality when certain ocular or systemic co-morbidities exist. Artifacts may result in inaccuracy or ambiguity in clinical diagnoses. In this paper, we proposed a simple but effective end-to-end framework for enhancing poor-quality retinal fundus images. Leveraging the optimal transport theory, we proposed an unpaired image-to-image translation scheme for transporting low-quality images to their high-quality counterparts. We theoretically proved that a Generative Adversarial Networks (GAN) model with a generator and discriminator is sufficient for this task. Furthermore, to mitigate the inconsistency of information between the low-quality images and their enhancements, an information consistency mechanism was proposed to maximally maintain structural consistency (optical discs, blood vessels, lesions) between the source and enhanced domains. Extensive experiments were conducted on the EyeQ dataset to demonstrate the superiority of our proposed method perceptually and quantitatively.
翻译:在存在某些眼部或系统的共发性时,摄影容易出现手工艺品、不完善和低质量,人工行为可能导致临床诊断不准确或模糊;在本文中,我们提议了一个简单而有效的端对端框架,用于提高低质量的视网膜基金图像;利用最佳运输理论,我们提议了一个将低质量图像传送到高品质的图像到图像翻译计划;我们理论上证明,配有生成器和制导器的基因反向网络(GAN)模型足以完成这项任务;此外,为了减少低质量图像及其增强之间的信息不一致,我们提议了一个信息一致性机制,以最大限度地保持源与强化区域之间的结构一致性(光碟、血管、损伤)。我们在EyeQ数据集上进行了广泛的实验,以显示我们拟议方法在认知和数量上具有优势。