Conventional unsupervised multi-source domain adaptation (UMDA) methods assume all source domains can be accessed directly. This neglects the privacy-preserving policy, that is, all the data and computations must be kept decentralized. There exists three problems in this scenario: (1) Minimizing the domain distance requires the pairwise calculation of the data from source and target domains, which is not accessible. (2) The communication cost and privacy security limit the application of UMDA methods (e.g., the domain adversarial training). (3) Since users have no authority to check the data quality, the irrelevant or malicious source domains are more likely to appear, which causes negative transfer. In this study, we propose a privacy-preserving UMDA paradigm named Knowledge Distillation based Decentralized Domain Adaptation (KD3A), which performs domain adaptation through the knowledge distillation on models from different source domains. KD3A solves the above problems with three components: (1) A multi-source knowledge distillation method named Knowledge Vote to learn high-quality domain consensus knowledge. (2) A dynamic weighting strategy named Consensus Focus to identify both the malicious and irrelevant domains. (3) A decentralized optimization strategy for domain distance named BatchNorm MMD. The extensive experiments on DomainNet demonstrate that KD3A is robust to the negative transfer and brings a 100x reduction of communication cost compared with other decentralized UMDA methods. Moreover, our KD3A significantly outperforms state-of-the-art UMDA approaches.
翻译:由于用户无权检查数据质量,因此不相干或恶意源域更有可能出现,从而导致负转移。在本研究中,我们提议一个名为“基于分散的Domain 适应(KD3A)的知识蒸馏”(KD3A)的维护隐私的UMDA模式,该模式通过对不同来源域的模型进行知识蒸馏来进行区域调整。KD3A解决上述问题有三个组成部分:(1) 名为“知识感化”的多源知识蒸馏方法,以学习高质量的域共识知识。(2) 名为“共识焦点”的动态加权战略,以识别恶意和不相关域,造成负转移。在这项研究中,我们提议一个名为“知识蒸馏”的UMDA范式模式(KD3A),以基于分散的Dmain适应(KDA3A)为基础,通过对不同源域域模型进行知识蒸馏来进行区域调整。KD3A节化战略以展示名为“MDA”的远程递增成本。