Despite achieving promising results in a breadth of medical image segmentation tasks, deep neural networks require large training datasets with pixel-wise annotations. Obtaining these curated datasets is a cumbersome process which limits the applicability in scenarios. Mixed supervision is an appealing alternative for mitigating this obstacle. In this work, we propose a dual-branch architecture, where the upper branch (teacher) receives strong annotations, while the bottom one (student) is driven by limited supervision and guided by the upper branch. Combined with a standard cross-entropy loss over the labeled pixels, our novel formulation integrates two important terms: (i) a Shannon entropy loss defined over the less-supervised images, which encourages confident student predictions in the bottom branch; and (ii) a KL divergence term, which transfers the knowledge (i.e., predictions) of the strongly supervised branch to the less-supervised branch and guides the entropy (student-confidence) term to avoid trivial solutions. We show that the synergy between the entropy and KL divergence yields substantial improvements in performance. We also discuss an interesting link between Shannon-entropy minimization and standard pseudo-mask generation, and argue that the former should be preferred over the latter for leveraging information from unlabeled pixels. We evaluate the effectiveness of the proposed formulation through a series of quantitative and qualitative experiments using two publicly available datasets. Results demonstrate that our method significantly outperforms other strategies for semantic segmentation within a mixed-supervision framework, as well as recent semi-supervised approaches. Our code is publicly available: https://github.com/by-liu/ConfKD.
翻译:尽管在广泛的医疗图像分割任务中取得了有希望的成果,但深神经网络需要大量的培训数据集,并配有像素说明。获取这些经整理的数据集是一个繁琐的过程,限制了在情景中的可应用性。混合监督是缓解这一障碍的一个诱人的选择。在这项工作中,我们提议一个双部门架构,由上部(教师)获得强有力的说明,而下部(学生)则受有限的监督,并由上部部门指导。加上在标签的像素上,我们的新配方包括了标准的跨性器官损失。我们的新配方结合了两个重要术语:(一) 香农分流分流损失,在不那么高的图像上部图像上下定义,鼓励学生作出自信的预测;以及(二) KL 差异术语,将严格监管的分支的知识(即,预测) 转移到不那么高的分支, 并指导昆虫(测试-信心) 术语,以避免微不足道的解决方案。我们显示, 英基和KL值之间的协同效应,在不高层次图像图像中带来显著的改进。我们还讨论一个令人感兴趣的链接- 标准化的版本的版本,通过前版数据流化的版本,通过前版数据流数据流数据流数据流数据流数据流数据流数据流,通过前版本,将我们现有的数据流数据流数据流数据流数据流数据流数据流数据流数据流数据流的版本,来解释。