In semantic segmentation, training data down-sampling is commonly performed because of limited resources, adapting image size to the model input, or improving data augmentation. This down-sampling typically employs different strategies for the image data and the annotated labels. Such discrepancy leads to mismatches between the down-sampled pixels and labels. Hence, training performance significantly decreases as the down-sampling factor increases. In this paper, we bring together the downsampling strategies for the image data and annotated labels. To that aim, we propose a soft-labeling method for label down-sampling that takes advantage of structural content prior to down-sampling. Thereby, fully aligning softlabels with image data to keep the distribution of the sampled pixels. This proposal also produces richer annotations for under-represented semantic classes. Altogether, it permits training competitive models at lower resolutions. Experiments show that the proposal outperforms other downsampling strategies. Moreover, state of the art performance is achieved for reference benchmarks, but employing significantly less computational resources than other approaches. This proposal enables competitive research for semantic segmentation under resource constraints.
翻译:在语义分解中,培训数据向下取样通常是由于资源有限、根据模型输入调整图像大小或改进数据扩增。这种向下取样通常对图像数据和附加说明的标签采用不同的策略。这种差异导致下抽样像素和标签之间的不匹配。因此,随着下抽样像素和标签的增加,培训性能显著下降。在本文中,我们将图像数据和附加说明的标签的向下取样战略汇集在一起。为了达到这个目的,我们提出了一种使用软标签向下取样的标签向下取样的方法,在下取样之前利用结构内容。因此,将软标签与图像数据完全匹配,以保持抽样像素的分布。这个建议还给代表不足的象素和标签的分布提供了更丰富的说明。总体来说,它允许在较低分辨率上培训竞争性模型。实验表明,该提议与其他下抽样战略相悖。此外,在参考基准方面实现了艺术性表现状况,但比其他方法下的计算资源使用要少得多。</s>