Annotating medical images for disease detection is often tedious and expensive. Moreover, the available training samples for a given task are generally scarce and imbalanced. These conditions are not conducive for learning effective deep neural models. Hence, it is common to 'transfer' neural networks trained on natural images to the medical image domain. However, this paradigm lacks in performance due to the large domain gap between the natural and medical image data. To address that, we propose a novel concept of Pre-text Representation Transfer (PRT). In contrast to the conventional transfer learning, which fine-tunes a source model after replacing its classification layers, PRT retains the original classification layers and updates the representation layers through an unsupervised pre-text task. The task is performed with (original, not synthetic) medical images, without utilizing any annotations. This enables representation transfer with a large amount of training data. This high-fidelity representation transfer allows us to use the resulting model as a more effective feature extractor. Moreover, we can also subsequently perform the traditional transfer learning with this model. We devise a collaborative representation based classification layer for the case when we leverage the model as a feature extractor. We fuse the output of this layer with the predictions of a model induced with the traditional transfer learning performed over our pre-text transferred model. The utility of our technique for limited and imbalanced data classification problem is demonstrated with an extensive five-fold evaluation for three large-scale models, tested for five different class-imbalance ratios for CT based COVID-19 detection. Our results show a consistent gain over the conventional transfer learning with the proposed method.
翻译:用于检测疾病的医疗图象往往乏味和昂贵。此外,用于特定任务的现有培训样本通常很少,而且不平衡。这些条件不利于学习有效的深神经模型。因此,通过自然图象培训的神经网络“转移”自然图象领域很常见。然而,由于自然和医学图象数据之间存在巨大的领域差距,这种模式在性能上缺乏。为了解决这个问题,我们提出了一个新的文本代表传输模式概念(PRT) 。与传统传输比例学习相比,这种传统传输比例在替换分类层后微调一个源模型,PRT保留原有的分类层,并通过未经监督的预文本任务更新代表层。这项任务是用(原始的而不是合成的)自然图象网络执行的。由于自然和医学图象数据之间的巨大领域差距,这种模式在性能上缺乏效果。为了使用传统的传输模型进行传统的传输模型,我们用基于传统格式的分类方法设计了一个基于合作的分类分类,我们用传统的分类方法更新了该模型的分类结果,我们用传统的版本模型来测试了一种不同的分类结果。 我们用一种不同的分类模型来测试了一种不同的分类图象变换数据。