How deep neural networks (DNNs) learn from noisy labels has been studied extensively in image classification but much less in image segmentation. So far, our understanding of the learning behavior of DNNs trained by noisy segmentation labels remains limited. In this study, we address this deficiency in both binary segmentation of biological microscopy images and multi-class segmentation of natural images. We generate extremely noisy labels by randomly sampling a small fraction (e.g., 10%) or flipping a large fraction (e.g., 90%) of the ground truth labels. When trained with these noisy labels, DNNs provide largely the same segmentation performance as trained by the original ground truth. This indicates that DNNs learn structures hidden in labels rather than pixel-level labels per se in their supervised training for semantic segmentation. We refer to these hidden structures in labels as meta-structures. When DNNs are trained by labels with different perturbations to the meta-structure, we find consistent degradation in their segmentation performance. In contrast, incorporation of meta-structure information substantially improves performance of an unsupervised segmentation model developed for binary semantic segmentation. We define meta-structures mathematically as spatial density distributions and show both theoretically and experimentally how this formulation explains key observed learning behavior of DNNs.
翻译:在图像分类中广泛研究了深层神经网络(DNN)如何从噪音标签中学习,但在图像分解方面却少得多。 到目前为止,我们对DNN通过噪音分解标签培训的学习行为的理解仍然有限。 在本研究中,我们解决生物显微镜图像二进制分解和自然图像多分类分解两方面的缺陷。我们通过随机抽样小部分(例如10%)或翻转地面真相标签的一大部分(例如90%)来产生极为吵动的标签。在接受这些吵闹标签培训时,DNNN提供与原始地面真理培训的大致相同的分解性能。这表明DNNNM学习了在标签中隐藏的结构,而不是象素级标签,而自然图像的多级分解。我们在标签中将这些隐藏的结构称为元结构。当DNNN受到与元结构不同的分解(例如90%)培训时,我们发现其分解性表现是一贯的。相比之下,将MDNNNNM信息融入到最初的数学分解模式中,我们从实质上解释了数学分解的数学分解过程。