Unsupervised domain adaptation (UDA) with pre-trained language models (PrLM) has achieved promising results since these pre-trained models embed generic knowledge learned from various domains. However, fine-tuning all the parameters of the PrLM on a small domain-specific corpus distort the learned generic knowledge, and it is also expensive to deployment a whole fine-tuned PrLM for each domain. This paper explores an adapter-based fine-tuning approach for unsupervised domain adaptation. Specifically, several trainable adapter modules are inserted in a PrLM, and the embedded generic knowledge is preserved by fixing the parameters of the original PrLM at fine-tuning. A domain-fusion scheme is introduced to train these adapters using a mix-domain corpus to better capture transferable features. Elaborated experiments on two benchmark datasets are carried out, and the results demonstrate that our approach is effective with different tasks, dataset sizes, and domain similarities.
翻译:未受监督的域适应(UDA)具有预先培训的语言模型(PrLM),这些经过预先培训的模型吸收了从不同领域获得的通用知识,因此取得了大有希望的成果。然而,微调PrLM关于一个小领域特有内容的所有参数,扭曲了所学的通用知识,对每个领域部署一个全精细调整的PrLM(UDA)也很昂贵。本文探讨了一个基于适应器的微调方法,用于未经监督的域适应。具体地说,几个可培训的适应器模块被插入了PrLM(PrLM)中,而嵌入的通用知识通过微调确定原PrLM(PrLM)的参数而得以保存。引入了一种域融合计划,用混合域特质来培训这些适应器,以更好地捕捉可转让的特性。对两个基准数据集进行了精心设计的实验,结果表明,我们的方法与不同的任务、数据集大小和领域相似性是有效的。