We consider the unsupervised scene adaptation problem of learning from both labeled source data and unlabeled target data. Existing methods focus on minoring the inter-domain gap between the source and target domains. However, the intra-domain knowledge and inherent uncertainty learned by the network are under-explored. In this paper, we propose an orthogonal method, called memory regularization in vivo to exploit the intra-domain knowledge and regularize the model training. Specifically, we refer to the segmentation model itself as the memory module, and minor the discrepancy of the two classifiers, i.e., the primary classifier and the auxiliary classifier, to reduce the prediction inconsistency. Without extra parameters, the proposed method is complementary to the most existing domain adaptation methods and could generally improve the performance of existing methods. Albeit simple, we verify the effectiveness of memory regularization on two synthetic-to-real benchmarks: GTA5 -> Cityscapes and SYNTHIA -> Cityscapes, yielding +11.1% and +11.3% mIoU improvement over the baseline model, respectively. Besides, a similar +12.0% mIoU improvement is observed on the cross-city benchmark: Cityscapes -> Oxford RobotCar.
翻译:我们认为,从标签源数据和未标签目标数据中学习的未经监督的场景适应性问题。现有方法侧重于缩小源与目标领域之间的差距。然而,网络所学到的内部知识和内在不确定性尚未得到充分探讨。在本文中,我们建议采用一个正统方法,称为体内记忆正规化,以利用内部知识并规范模式培训。具体地说,我们把分割模型本身称为存储模块,并略微缩小两个分类器的差异,即主要分类器和辅助分类器,以减少预测不一致之处。没有额外的参数,拟议方法就是对现有大多数领域适应方法的补充,并一般可以改进现有方法的性能。尽管简单,我们核查两个合成到现实基准上的记忆正规化的有效性:GTA5 - > 城景和SYNTHIA - > 城景,在基线模型中分别产生+11.1%和+11.3% mOU改进。此外,一个类似的基准是城市+12 % 。我们核查了两个合成到的城市的升级率。