Computed tomography (CT) imaging is a promising approach to diagnosing the COVID-19. Machine learning methods can be employed to train models from labeled CT images and predict whether a case is positive or negative. However, there exists no publicly-available and large-scale CT data to train accurate models. In this work, we propose a multi-stage attentive transfer learning framework for improving COVID-19 diagnosis. Our proposed framework consists of three stages to train accurate diagnosis models through learning knowledge from multiple source tasks and data of different domains. Importantly, we propose a novel self-supervised learning method to learn multi-scale representations for lung CT images. Our method captures semantic information from the whole lung and highlights the functionality of each lung region for better representation learning. The method is then integrated to the last stage of the proposed transfer learning framework to reuse the complex patterns learned from the same CT images. We use a base model integrating self-attention (ATTNs) and convolutional operations. Experimental results show that networks with ATTNs induce greater performance improvement through transfer learning than networks without ATTNs. This indicates attention exhibits higher transferability than convolution. Our results also show that the proposed self-supervised learning method outperforms several baseline methods.
翻译:计算机学习方法可用于从贴标签的CT图像中培训模型,并预测一个病例是正的还是负的。但是,没有公开和大规模的CT数据来培训准确的模型。在这项工作中,我们提议了一个多阶段的注意转移学习框架,以改进COVID-19的诊断。我们提议的框架包括三个阶段,通过从多种来源任务和不同领域的数据学习知识来培训准确的诊断模型。重要的是,我们提出了一种新的自我监督学习方法,以学习肺部CT图像的多尺度表现方式。我们的方法收集了整个肺部的语义信息,并突出了每个肺部区域的功能,以更好地进行陈述学习。然后,我们将这一方法整合到拟议的转移学习框架的最后阶段,以重新利用从相同的CT图像中学习的复杂模式。我们使用一个基础模型,将自留(ATNs)和进化操作结合起来。实验结果显示,与ATTNs建立的网络通过转移学习比没有ATTNS图像的多尺度表现改进了绩效。我们的方法从若干项基线上展示了自我转移的方法。