Semantic image segmentation is a central and challenging task in autonomous driving, addressed by training deep models. Since this training draws to a curse of human-based image labeling, using synthetic images with automatically generated labels together with unlabeled real-world images is a promising alternative. This implies to address an unsupervised domain adaptation (UDA) problem. In this paper, we propose a new co-training procedure for synth-to-real UDA of semantic segmentation models. It consists of a self-training stage, which provides two domain-adapted models, and a model collaboration loop for the mutual improvement of these two models. These models are then used to provide the final semantic segmentation labels (pseudo-labels) for the real-world images. The overall procedure treats the deep models as black boxes and drives their collaboration at the level of pseudo-labeled target images, i.e., neither modifying loss functions is required, nor explicit feature alignment. We test our proposal on standard synthetic and real-world datasets for on-board semantic segmentation. Our procedure shows improvements ranging from ~13 to ~26 mIoU points over baselines, so establishing new state-of-the-art results.
翻译:语义图像分解是自主驱动中一项核心和具有挑战性的任务,通过培养深层模型加以解决。 由于这一培训将人类图像标签的诅咒引向基于人类图像标签的诅咒, 使用自动生成标签的合成图像和未贴标签的现实世界图像是一个很有希望的替代办法。 这意味着要解决一个不受监督的域适应问题。 在本文件中, 我们为语义分解模型的合成到真实 UDA提出了一个新的联合培训程序。 它包括一个自我培训阶段, 提供两种域适应模型, 以及用于两套模型相互改进的模型合作循环模式。 这些模型随后被用来为真实世界图像提供最后的语义分解标签( 假名标签 ) 。 总体程序将深层模型作为黑盒, 推动其在假标签目标图像层面的合作, 即, 不需要修改损失函数, 也不需要明确的特征校正。 我们测试了我们关于标准合成和真实世界数据集成的模板, 用于机体语义分解的模型。 我们的程序显示从 m13 到 m- 基准点的改进范围 。