Many unsupervised domain adaptation (UDA) methods exploit domain adversarial training to align the features to reduce domain gap, where a feature extractor is trained to fool a domain discriminator in order to have aligned feature distributions. The discrimination capability of the domain classifier w.r.t the increasingly aligned feature distributions deteriorates as training goes on, thus cannot effectively further drive the training of feature extractor. In this work, we propose an efficient optimization strategy named Re-enforceable Adversarial Domain Adaptation (RADA) which aims to re-energize the domain discriminator during the training by using dynamic domain labels. Particularly, we relabel the well aligned target domain samples as source domain samples on the fly. Such relabeling makes the less separable distributions more separable, and thus leads to a more powerful domain classifier w.r.t. the new data distributions, which in turn further drives feature alignment. Extensive experiments on multiple UDA benchmarks demonstrate the effectiveness and superiority of our RADA.
翻译:许多未经监督的域适应方法(UDA)利用域对称培训来调整功能以缩小域间差距,其中地物提取器经过训练以愚弄一个域歧视者,以便取得一致的地物分布。随着培训的继续,域分类器(w.r.t)的特性分布越来越一致,因此无法有效地进一步推动地物提取器的培训。在这项工作中,我们提议了一个有效的优化战略,名为“重新加强可执行的Aversarial Domain适应”(RADA),目的是在培训期间通过动态域名标签重新激活域内区分器(RADA)。特别是,我们把目标相匹配的域样重新标为飞的源域域样。这种重新标签使较不易分离的地物分布更加容易分离,从而导致更强大的域域分类器(w.r.t.)新的数据分布,反过来又使驱动器特性更加一致。关于多种UDADA基准的广泛实验显示了我们的功效和优越性。