Domain adaptation has attracted a great deal of attention in the machine learning community, but it requires access to source data, which often raises concerns about data privacy. We are thus motivated to address these issues and propose a simple yet efficient method. This work treats domain adaptation as an unsupervised clustering problem and trains the target model without access to the source data. Specifically, we propose a loss function called contrast and clustering (CaC), where a positive pair term pulls neighbors belonging to the same class together in the feature space to form clusters, while a negative pair term pushes samples of different classes apart. In addition, extended neighbors are taken into account by querying the nearest neighbor indexes in the memory bank to mine for more valuable negative pairs. Extensive experiments on three common benchmarks, VisDA, Office-Home and Office-31, demonstrate that our method achieves state-of-the-art performance. The code will be made publicly available at https://github.com/yukilulu/CaC.
翻译:在机器学习界,对域的适应引起了极大关注,但需要获取源数据,这往往引起对数据隐私的关切。因此,我们有动力解决这些问题并提出简单而有效的方法。这项工作将域适应视为一个不受监督的集群问题,并培训了目标模型,而没有获得源数据。具体地说,我们提议了一个称为对比和集群(CaC)的损失函数,其中正对术语将属于同一类的邻居聚集在地物空间,形成群落,而负对称术语将不同类别的样本分开。此外,通过查询记忆库中最近的近邻索引,让更有价值的负对子进入矿井,也考虑到了扩大的邻居。关于三个共同基准(VisDA、Office-Home和Office-31)的广泛实验表明,我们的方法达到了最新技术性能。该代码将在https://github.com/yukilulu/C上公布。