Deep Learning has greatly advanced the performance of semantic segmentation, however, its success relies on the availability of large amounts of annotated data for training. Hence, many efforts have been devoted to domain adaptive semantic segmentation that focuses on transferring semantic knowledge from a labeled source domain to an unlabeled target domain. Existing self-training methods typically require multiple rounds of training, while another popular framework based on adversarial training is known to be sensitive to hyper-parameters. In this paper, we present an easy-to-train framework that learns domain-invariant prototypes for domain adaptive semantic segmentation. In particular, we show that domain adaptation shares a common character with few-shot learning in that both aim to recognize some types of unseen data with knowledge learned from large amounts of seen data. Thus, we propose a unified framework for domain adaptation and few-shot learning. The core idea is to use the class prototypes extracted from few-shot annotated target images to classify pixels of both source images and target images. Our method involves only one-stage training and does not need to be trained on large-scale un-annotated target images. Moreover, our method can be extended to variants of both domain adaptation and few-shot learning. Experiments on adapting GTA5-to-Cityscapes and SYNTHIA-to-Cityscapes show that our method achieves competitive performance to state-of-the-art.
翻译:深层学习极大地提高了语义分解的性能,然而,它的成功取决于提供大量附加说明的培训数据。因此,许多努力都致力于领域适应性语义分解,重点是将语义学知识从标签源域向无标签目标域转移。现有的自我培训方法通常需要多轮培训,而另一个基于对抗性培训的流行框架已知对超参数十分敏感。在本文中,我们提出了一个容易到火车的框架,用于学习领域适应性语义分解的域异变原型。特别是,我们展示了域适应性语言分解有一个共同的共性,其中少有的学习,目的是承认从标签源域域向无标签目标域传授知识的某类隐性数据。因此,我们提出了一套统一的域性适应框架和几张光学学习。核心理念是使用从少发附加说明性目标图像中提取的班级原型模型,对源图像和目标图像进行分类。我们的方法仅涉及一阶段培训,不需要在大规模C级变换目标图像上进行相同的学习。此外,我们的方法可以扩展为大规模C级变式变换GTIS图。