Recently, deep-learning-based approaches have been widely studied for deformable image registration task. However, most efforts directly map the composite image representation to spatial transformation through the convolutional neural network, ignoring its limited ability to capture spatial correspondence. On the other hand, Transformer can better characterize the spatial relationship with attention mechanism, its long-range dependency may be harmful to the registration task, where voxels with too large distances are unlikely to be corresponding pairs. In this study, we propose a novel Deformer module along with a multi-scale framework for the deformable image registration task. The Deformer module is designed to facilitate the mapping from image representation to spatial transformation by formulating the displacement vector prediction as the weighted summation of several bases. With the multi-scale framework to predict the displacement fields in a coarse-to-fine manner, superior performance can be achieved compared with traditional and learning-based approaches. Comprehensive experiments on two public datasets are conducted to demonstrate the effectiveness of the proposed Deformer module as well as the multi-scale framework.
翻译:最近,对基于深层学习的方法进行了广泛的研究,以完成可变图像登记任务;然而,大多数努力都直接通过进化神经网络绘制综合图像显示空间转换图,而忽略其捕捉空间通信的有限能力;另一方面,变异器可以更好地描述空间关系与注意机制的特征,其远距离依赖性可能有害于登记任务,因为在这种任务中,距离过远的氧化物不大可能是对应的对应配对;在本研究中,我们提议建立一个新型变异模块,同时提出可变异图像登记任务的多尺度框架;变异模块旨在通过将移位矢量预测作为若干基数的加权总和,促进从图像显示到空间转换的绘图;随着以粗到细的方式预测迁移场的多尺度框架,与传统和以学习为基础的方法相比,可以取得更高的性能;对两个公共数据集进行了全面试验,以证明拟议的变异模块和多尺度框架的有效性。