In this paper, we seek reasons for the two major failure cases in Semantic Segmentation (SS): 1) missing small objects or minor object parts, and 2) mislabeling minor parts of large objects as wrong classes. We have an interesting finding that Failure-1 is due to the underuse of detailed features and Failure-2 is due to the underuse of visual contexts. To help the model learn a better trade-off, we introduce several Self-Regulation (SR) losses for training SS neural networks. By "self", we mean that the losses are from the model per se without using any additional data or supervision. By applying the SR losses, the deep layer features are regulated by the shallow ones to preserve more details; meanwhile, shallow layer classification logits are regulated by the deep ones to capture more semantics. We conduct extensive experiments on both weakly and fully supervised SS tasks, and the results show that our approach consistently surpasses the baselines. We also validate that SR losses are easy to implement in various state-of-the-art SS models, e.g., SPGNet and OCRNet, incurring little computational overhead during training and none for testing.
翻译:在本文中,我们寻求在语义分割(SS)中出现两大故障案例的原因:(1) 小物体或次要物体部件失踪,(2) 大物体的细小部分被误标为错误类别。我们有一个有趣的发现,失败1号是由于没有充分利用详细特性,失败2号是由于没有充分利用视觉环境。为了帮助模型学习更好的权衡,我们为培训SS神经网络引入了一些自我调节(SR)损失。“自我”是指损失来自模型本身,而没有使用任何额外数据或监督。通过应用SR损失,深层特征由浅层特征管理,以保存更多细节;与此同时,浅层分类日志由深层分类日志管理,以捕捉更多的语义。我们对于薄弱和完全受监管的SS任务进行了广泛的实验,结果显示我们的方法始终超过基线。我们还证实,SR损失很容易在各种状态的SS模型中实施,例如SPGNet和OCRNet,在训练期间很少进行计算,没有用于测试。