Vision foundation models have demonstrated strong generalization in medical image segmentation by leveraging large-scale, heterogeneous pretraining. However, they often struggle to generalize to specialized clinical tasks under limited annotations or rare pathological variations, due to a mismatch between general priors and task-specific requirements. To address this, we propose Uncertainty-informed Collaborative Learning (UnCoL), a dual-teacher framework that harmonizes generalization and specialization in semi-supervised medical image segmentation. Specifically, UnCoL distills both visual and semantic representations from a frozen foundation model to transfer general knowledge, while concurrently maintaining a progressively adapting teacher to capture fine-grained and task-specific representations. To balance guidance from both teachers, pseudo-label learning in UnCoL is adaptively regulated by predictive uncertainty, which selectively suppresses unreliable supervision and stabilizes learning in ambiguous regions. Experiments on diverse 2D and 3D segmentation benchmarks show that UnCoL consistently outperforms state-of-the-art semi-supervised methods and foundation model baselines. Moreover, our model delivers near fully supervised performance with markedly reduced annotation requirements.
翻译:视觉基础模型通过大规模、异构的预训练在医学图像分割中展现出强大的泛化能力。然而,由于通用先验与任务特定需求之间的不匹配,这些模型在标注有限或罕见病理变异下的专业化临床任务中往往难以有效泛化。为解决这一问题,我们提出不确定性协同学习(UnCoL),一种双教师框架,旨在半监督医学图像分割中协调泛化与特化。具体而言,UnCoL从冻结的基础模型中蒸馏视觉和语义表示以传递通用知识,同时维护一个逐步自适应的教师模型以捕获细粒度的任务特定表示。为平衡两位教师的指导,UnCoL中的伪标签学习通过预测不确定性进行自适应调节,选择性地抑制不可靠的监督并稳定模糊区域的学习。在多种2D和3D分割基准上的实验表明,UnCoL持续优于当前最先进的半监督方法和基础模型基线。此外,我们的模型在显著降低标注需求的情况下实现了接近全监督的性能。