Traditional supervised learning with deep neural networks requires a tremendous amount of labelled data to converge to a good solution. For 3D medical images, it is often impractical to build a large homogeneous annotated dataset for a specific pathology. Self-supervised methods offer a new way to learn a representation of the images in an unsupervised manner with a neural network. In particular, contrastive learning has shown great promises by (almost) matching the performance of fully-supervised CNN on vision tasks. Nonetheless, this method does not take advantage of available meta-data, such as participant's age, viewed as prior knowledge. Here, we propose to leverage continuous proxy metadata, in the contrastive learning framework, by introducing a new loss called y-Aware InfoNCE loss. Specifically, we improve the positive sampling during pre-training by adding more positive examples with similar proxy meta-data with the anchor, assuming they share similar discriminative semantic features.With our method, a 3D CNN model pre-trained on $10^4$ multi-site healthy brain MRI scans can extract relevant features for three classification tasks: schizophrenia, bipolar diagnosis and Alzheimer's detection. When fine-tuned, it also outperforms 3D CNN trained from scratch on these tasks, as well as state-of-the-art self-supervised methods. Our code is made publicly available here.
翻译:由深层神经网络监督的传统学习需要大量的贴标签数据才能融合到一个良好的解决方案。 对于 3D 医疗图像来说, 为特定病理学建立一个庞大的同质附加说明数据集往往不切实际。 自监督的方法为以不受监督的方式与神经网络学习图像的描述提供了一种新的方式。 特别是, 对比式学习显示了巨大的承诺, 因为它( 几乎) 与完全监督的CNN 的功能匹配了完全监督的视觉任务。 然而, 这种方法并没有利用现有的元数据, 如参与者的年龄, 被视为先前的知识。 在这里, 我们提议在对比性学习框架内利用连续的代理元数据, 引入一个新的损失叫做 y-Aware InfoNCE 损失 。 具体地说, 我们通过在培训前的测试中添加更正面的例子, 类似的代理元数据与锚具有相似的区别性语言特征。 用我们的方法, 3DCNNCNMMMMM 进行多点健康脑扫描, 能够为三种分类任务提取相关的特征: Schizzo-D- droad of the gregroduction of the sheal-