In ophthalmological imaging, multiple imaging systems, such as color fundus, infrared, fluorescein angiography, optical coherence tomography (OCT) or OCT angiography, are often involved to make a diagnosis of retinal disease. Multi-modal retinal registration techniques can assist ophthalmologists by providing a pixel-based comparison of aligned vessel structures in images from different modalities or acquisition times. To this end, we propose an end-to-end trainable deep learning method for multi-modal retinal image registration. Our method extracts convolutional features from the vessel structure for keypoint detection and description and uses a graph neural network for feature matching. The keypoint detection and description network and graph neural network are jointly trained in a self-supervised manner using synthetic multi-modal image pairs and are guided by synthetically sampled ground truth homographies. Our method demonstrates higher registration accuracy as competing methods for our synthetic retinal dataset and generalizes well for our real macula dataset and a public fundus dataset.
翻译:在眼科成像中,多种成像系统,如彩球、红外线、荧光血管成像仪、光一致性成像仪或OCT血管成像仪,常常参与诊断视网膜疾病;多式视网膜登记技术可以帮助眼科医生,对不同方式或获取时间的图像中相配的船舶结构进行像素比较;为此,我们提议了一种从终端到终端的可训练的多式视网膜图像登记深层学习方法。我们的方法从船舶结构中提取关键点探测和描述的脉冲特征,并使用图形神经网络进行地貌匹配。关键点探测和描述网络和图形神经网络,使用合成多式多式图像配对,以自我监督的方式共同培训,并以合成抽样的地面真象同质谱系统为指导。我们的方法显示较高的登记准确性,作为我们合成视网膜数据集的相互竞争方法,以及用于我们真实的数学数据集和公共基金数据集的普通化方法。