Despite advances in data augmentation and transfer learning, convolutional neural networks (CNNs) difficultly generalise to unseen domains. When segmenting brain scans, CNNs are highly sensitive to changes in resolution and contrast: even within the same MRI modality, performance can decrease across datasets. Here we introduce SynthSeg, the first segmentation CNN robust against changes in contrast and resolution. SynthSeg is trained with synthetic data sampled from a generative model conditioned on segmentations. Crucially, we adopt a domain randomisation strategy where we fully randomise the contrast and resolution of the synthetic training data. Consequently, SynthSeg can segment real scans from a wide range of target domains without retraining or fine-tuning, which enables straightforward analysis of huge amounts of heterogeneous clinical data. Because SynthSeg only requires segmentations to be trained (no images), it can learn from labels obtained by automated methods on diverse populations (e.g., ageing and diseased), thus achieving robustness to a wide range of morphological variability. We demonstrate SynthSeg on 5,000 scans of six modalities (including CT) and ten resolutions, where it exhibits unparalleled generalisation compared with supervised CNNs, state-of-the-art domain adaptation, and Bayesian segmentation. Finally, we demonstrate the generalisability of SynthSeg by applying it to cardiac MRI and CT scans.
翻译:尽管在数据增加和传输学习方面有所进步,但进化神经网络(CNNs)难以向隐蔽领域推广。在对大脑进行分解扫描时,CNN对分辨率和对比的变化非常敏感:即使在同一MRI模式下,性能也可以降低跨数据集的功能。在这里,我们引入了SynthSeg,即第一个分解CNNEG, 对抗对比和分辨率变化的强力SynthSeg。SynthSeg是用以分化为条件的基因模型对合成数据进行抽样抽样的合成数据培训的。关键是,我们采用一种域随机化战略,完全随机化合成培训数据的对比和解析。因此,SynthSeg可以将一系列目标领域的真实扫描进行分解,而不进行再培训或微调,从而能够直接分析大量的差异性临床数据。由于SynthSegg只要求进行分解(无图像),因此它可以学习以自动化方法获得的关于不同人群(如老龄化和疾病等)的标签,从而实现广泛的形态变异性。我们展示了5 000次的扫描(包括CT)和CMSICNS总分辨率的升级,最后展示了全域。