Due to the scarcity of dense pixel-level semantic annotations for images recorded in adverse visual conditions, there has been a keen interest in unsupervised domain adaptation (UDA) for the semantic segmentation of such images. UDA adapts models trained on normal conditions to the target adverse-condition domains. Meanwhile, multiple datasets with driving scenes provide corresponding images of the same scenes across multiple conditions, which can serve as a form of weak supervision for domain adaptation. We propose Refign, a generic extension to self-training-based UDA methods which leverages these cross-domain correspondences. Refign consists of two steps: (1) aligning the normal-condition image to the corresponding adverse-condition image using an uncertainty-aware dense matching network, and (2) refining the adverse prediction with the normal prediction using an adaptive label correction mechanism. We design custom modules to streamline both steps and set the new state of the art for domain-adaptive semantic segmentation on several adverse-condition benchmarks, including ACDC and Dark Zurich. The approach introduces no extra training parameters, minimal computational overhead -- during training only -- and can be used as a drop-in extension to improve any given self-training-based UDA method. Code is available at https://github.com/brdav/refign.
翻译:由于对在不利视觉条件下录制的图像缺乏密集的像素级语义说明,因此对未经监督的域适应(UDA)对此类图像的语义分解非常感兴趣。UDA将通常条件经过培训的模型适应目标的不利条件域;同时,多个带有驱动场景的数据集提供多个条件下相同场景的相应图像,这可以作为对域适应的薄弱监督的一种形式。我们提议对基于自我训练的UDA方法进行回调,对基于自我训练的UCDC和黑暗苏黎世等一些不利条件基准进行通用扩展。重新定位包括两个步骤:(1)将正常条件图像与相应的不利条件图像对齐,使用不确定性-觉悟密集的匹配网络,(2)使用适应标签校正机制,用正常预测改进负面预测。我们设计定制模块,以精简两个步骤,并设定对域适应性地震分解的新状态,这些基准包括ACDC和黑暗苏黎世。该方法没有引入额外的培训参数,仅使用基于最小的计算间接间接 -- -- 仅在培训期间使用 -- -- 并且可以用作自动测试/调制。