This paper studies 3D low-dose computed tomography (CT) imaging. Although various deep learning methods were developed in this context, typically they perform denoising due to low-dose and deblurring for super-resolution separately. Up to date, little work was done for simultaneous in-plane denoising and through-plane deblurring, which is important to improve clinical CT images. For this task, a straightforward method is to directly train an end-to-end 3D network. However, it demands much more training data and expensive computational costs. Here, we propose to link in-plane and through-plane transformers for simultaneous in-plane denoising and through-plane deblurring, termed as LIT-Former, which can efficiently synergize in-plane and through-plane sub-tasks for 3D CT imaging and enjoy the advantages of both convolution and transformer networks. LIT-Former has two novel designs: efficient multi-head self-attention modules (eMSM) and efficient convolutional feed-forward networks (eCFN). First, eMSM integrates in-plane 2D self-attention and through-plane 1D self-attention to efficiently capture global interactions of 3D self-attention, the core unit of transformer networks. Second, eCFN integrates 2D convolution and 1D convolution to extract local information of 3D convolution in the same fashion. As a result, the proposed LIT-Former synergizes these two sub-tasks, significantly reducing the computational complexity as compared to 3D counterparts and enabling rapid convergence. Extensive experimental results on simulated and clinical datasets demonstrate superior performance over state-of-the-art models.
翻译:本文研究的是 3D 低剂量计算成色成像(CT) 。 虽然在此背景下开发了各种深层次学习方法, 但通常由于低剂量和分流导致超分辨率的低剂量和分流而进行分解。 到目前为止, 用于同时在机内分解和通过平板分流的工作很少, 这对于改善临床CT 图像非常重要。 对于这项任务来说, 一个直截了当的方法是直接培训端至端3D 网络。 但是, 它需要两个新颖的设计: 高效的多头自控模块( EMM) 和高效的平流式变压变异器变异器, 被称为LIT- Formering( LIT-Forning) 的同时在机内分解和通过平流机拆分解的同时进行分解。 将3D 3D 核心成像和变异器网络的自我同步同步, 将这三代变异变变变变变码的自我演算结果通过1 CMIS 的自我演算, 将这三代的自我变变变变变变变变变变变的自我演变变变的自我演变变变变的自我结果 。