Benefiting from considerable pixel-level annotations collected from a specific situation (source), the trained semantic segmentation model performs quite well but fails in a new situation (target). To mitigate the domain gap, previous cross-domain semantic segmentation methods always assume the co-existence of source data and target data during domain alignment. However, accessing source data in the real scenario may raise privacy concerns and violate intellectual property. To tackle this problem, we focus on an interesting and challenging cross-domain semantic segmentation task where only the trained source model is provided to the target domain. Specifically, we propose a unified framework called \textbf{ATP}, which consists of three schemes, i.e., feature \textbf{A}lignment, bidirectional \textbf{T}eaching, and information \textbf{P}ropagation. First, considering explicit alignment is infeasible due to no source data, we devise a curriculum-style entropy minimization objective to implicitly align the target features with unseen source features via the provided source model. Second, besides positive pseudo labels in vanilla self-training, we introduce negative pseudo labels to this field and develop a bidirectional self-training strategy to enhance the representation learning in the target domain. It is the first work to use negative pseudo labels during self-training for domain adaptation. Finally, the information propagation scheme is employed to further reduce the intra-domain discrepancy within the target domain via pseudo-semi-supervised learning, which is the first step by providing a simple and effective post-process for the domain adaptation field. Furthermore, we also extend the proposed to the more challenging black-box source-model scenario where only the source model's prediction is available.
翻译:从从特定情况(来源)中收集的相当像素级别的批注中获益的,经过培训的语义分解模型运行良好,但在新情况(目标)中却失败。为了缩小域间差距,以前的跨部语义分解方法总是假设源数据和目标数据在域对齐期间共存。然而,在真实情况下访问源数据可能会引起隐私关切,并侵犯知识产权。为了解决这一问题,我们侧重于一个有趣的、具有挑战性的跨部语义分解任务,即只向目标域提供经过培训的源码模型。具体地说,我们提议了一个称为\ textbf{ATP}的统一框架,这个框架由三个方案组成,即功能\ textbf{A}光化、双向导\textb{T}示意、以及信息\textb{P}ropagation。为了解决这一问题,我们考虑明确的校正校正的校正校正校正校正,我们设计了一个课程式的最小化变现最小化目标功能,以便通过提供的源码模型模型模型将目标特性与隐性源码特性统一起来。第二,除了正的域域内自我校正的校正校正的校正的校正的校正的校正校正的校正的校正的校正的校正的校正的校正的校正的校正的校正的校正的校正的校正外校正的校正的校正外的校正外校正外的校正外的校正外校正外校正外校正外校正外校方程式,它。