Recently, deep learning with Convolutional Neural Networks (CNNs) and Transformers has shown encouraging results in fully supervised medical image segmentation. However, it is still challenging for them to achieve good performance with limited annotations for training. In this work, we present a very simple yet efficient framework for semi-supervised medical image segmentation by introducing the cross teaching between CNN and Transformer. Specifically, we simplify the classical deep co-training from consistency regularization to cross teaching, where the prediction of a network is used as the pseudo label to supervise the other network directly end-to-end. Considering the difference in learning paradigm between CNN and Transformer, we introduce the Cross Teaching between CNN and Transformer rather than just using CNNs. Experiments on a public benchmark show that our method outperforms eight existing semi-supervised learning methods just with a simpler framework. Notably, this work may be the first attempt to combine CNN and transformer for semi-supervised medical image segmentation and achieve promising results on a public benchmark. The code will be released at: https://github.com/HiLab-git/SSL4MIS.
翻译:最近,与进化神经网络(CNNs)和变异器的深层次学习显示,在完全监管的医疗图像分割方面,取得了令人鼓舞的成果;然而,对于他们来说,在培训说明有限的情况下,取得良好的业绩仍是一项挑战;在这项工作中,我们提出了一个非常简单而有效的半监管医疗图像分割框架,方法是引入CNN和变异器之间的交叉教学。具体地说,我们简化了传统的深层次共同培训,从一致性正规化到交叉教学,将网络的预测用作假标签,直接监督其他网络的终端到终端。考虑到CNN和变异器之间在学习模式上的差别,我们引入CNN和变异器之间的交叉教学,而不是仅仅使用CNNs。在公共基准上的实验显示,我们的方法比现有的八种半监管学习方法简单化。值得注意的是,这项工作可能是首次尝试将CNN和变异器结合起来,用于半超强的医疗图像分割,并在公共基准上取得有希望的结果。该代码将在以下网站发布:https://github.com/HiLab-gigit/SSL4MIS。