Semantic segmentation of nighttime images plays an equally important role as that of daytime images in autonomous driving, but the former is much more challenging due to poor illuminations and arduous human annotations. In this paper, we propose a novel domain adaptation network (DANNet) for nighttime semantic segmentation without using labeled nighttime image data. It employs an adversarial training with a labeled daytime dataset and an unlabeled dataset that contains coarsely aligned day-night image pairs. Specifically, for the unlabeled day-night image pairs, we use the pixel-level predictions of static object categories on a daytime image as a pseudo supervision to segment its counterpart nighttime image. We further design a re-weighting strategy to handle the inaccuracy caused by misalignment between day-night image pairs and wrong predictions of daytime images, as well as boost the prediction accuracy of small objects. The proposed DANNet is the first one stage adaptation framework for nighttime semantic segmentation, which does not train additional day-night image transfer models as a separate pre-processing stage. Extensive experiments on Dark Zurich and Nighttime Driving datasets show that our method achieves state-of-the-art performance for nighttime semantic segmentation.
翻译:夜间图像的语义分割与白天图像在自主驱动中扮演着同等重要的角色, 但前者则由于光亮度低和人文说明艰苦而更具挑战性。 在本文中, 我们提议在不使用贴标签的夜间图像数据的情况下, 为夜间语义分割建立一个新型域适应网络( DANNet) 。 它使用贴标签的日间数据集和含有粗糙对齐日间夜图像配对的未贴标签数据集的对抗性培训。 具体来说, 对于无标签的日间图像配对而言, 我们使用日间图像中静态物体类别的像素水平预测作为假监督来分割其对等夜间图像。 我们进一步设计了一种重新加权战略, 处理夜间图像配对配对与日间图像错误预测造成的不准确性, 并提升小天体的预测准确性。 拟议的丹网是夜间图像分割的第一个阶段性适应框架, 它不会将额外的夜间图像转换模型培训成一个单独的夜间图像传输模型, 以单独的深夜间分析阶段显示我们的黑暗区段 。