Deep learning models, e.g. supervised Encoder-Decoder style networks, exhibit promising performance in medical image segmentation, but come with a high labelling cost. We propose TriSegNet, a semi-supervised semantic segmentation framework. It uses triple-view feature learning on a limited amount of labelled data and a large amount of unlabeled data. The triple-view architecture consists of three pixel-level classifiers and a low-level shared-weight learning module. The model is first initialized with labelled data. Label processing, including data perturbation, confidence label voting and unconfident label detection for annotation, enables the model to train on labelled and unlabeled data simultaneously. The confidence of each model gets improved through the other two views of the feature learning. This process is repeated until each model reaches the same confidence level as its counterparts. This strategy enables triple-view learning of generic medical image datasets. Bespoke overlap-based and boundary-based loss functions are tailored to the different stages of the training. The segmentation results are evaluated on four publicly available benchmark datasets including Ultrasound, CT, MRI, and Histology images. Repeated experiments demonstrate the effectiveness of the proposed network compared against other semi-supervised algorithms, across a large set of evaluation measures.
翻译:深度学习模型,例如受监督的 Eccoder-Decoder-Decoder 风格网络,在医疗图像分割方面表现出有良好的业绩,但标签成本很高。我们提议TriSegNet,这是一个半受监督的语义分割框架。它使用三眼特征学习有限数量标签数据和大量未贴标签数据。三眼结构由三个像素等级分类和低级别共享学习模块组成。该模型首先由贴标签数据组成。Label处理,包括数据渗透、信任标签投票和识别不自信标签以备注解,使该模型能够同时培训贴标签和未贴标签的数据。每个模型的信心通过功能学习的其他两种观点得到提高。这一过程将重复到每个模型达到与对应方相同的信任水平。这个战略可以让通用医学图像数据集的三眼学习。根据重叠和基于边界的损失功能适应了培训的不同阶段。分解结果在四个公开的基准数据集上进行了评估,包括Ultrazurburation、Meximicalal 和跨其拟议的大规模图像的模拟测试。