The computer-aided disease diagnosis from radiomic data is important in many medical applications. However, developing such a technique relies on annotating radiological images, which is a time-consuming, labor-intensive, and expensive process. In this work, we present the first novel collaborative self-supervised learning method to solve the challenge of insufficient labeled radiomic data, whose characteristics are different from text and image data. To achieve this, we present two collaborative pretext tasks that explore the latent pathological or biological relationships between regions of interest and the similarity and dissimilarity information between subjects. Our method collaboratively learns the robust latent feature representations from radiomic data in a self-supervised manner to reduce human annotation efforts, which benefits the disease diagnosis. We compared our proposed method with other state-of-the-art self-supervised learning methods on a simulation study and two independent datasets. Extensive experimental results demonstrated that our method outperforms other self-supervised learning methods on both classification and regression tasks. With further refinement, our method shows the potential advantage in automatic disease diagnosis with large-scale unlabeled data available.
翻译:利用放射数据进行计算机辅助疾病诊断在许多医疗应用中非常重要。然而,开发这种技术依赖于批注放射图像,这是一个耗时、劳动密集型和昂贵的过程。在这项工作中,我们提出了第一种新型的协作性自我监督学习方法,以解决贴有标签的辐射数据不足的挑战,其特性不同于文字和图像数据。为了实现这一目标,我们提出了两个协作的借口任务,探索感兴趣的区域之间的潜在病理学或生物关系,以及不同对象之间的相似性和差异性信息。我们的方法通过自我监督的方式,从放射数据中学习了强大的潜在特征,以减少人类的注解工作,这有利于疾病诊断。我们在模拟研究和两个独立的数据集中比较了我们所提议的方法。广泛的实验结果表明,我们的方法超越了分类和回归任务方面其他自我监督的学习方法。通过进一步的改进,我们的方法显示了在自动疾病诊断方面的潜在优势,即可用大规模未贴标签的数据。