Pivot-based neural representation models have lead to significant progress in domain adaptation for NLP. However, previous works that follow this approach utilize only labeled data from the source domain and unlabeled data from the source and target domains, but neglect to incorporate massive unlabeled corpora that are not necessarily drawn from these domains. To alleviate this, we propose PERL: A representation learning model that extends contextualized word embedding models such as BERT with pivot-based fine-tuning. PERL outperforms strong baselines across 22 sentiment classification domain adaptation setups, improves in-domain model performance, yields effective reduced-size models and increases model stability.
翻译:以枢纽为主的神经代表模型导致在NLP领域适应方面取得重大进展。然而,以往采用这一方法的工作仅使用来源领域贴标签的数据和来源和目标领域未贴标签的数据,但忽视纳入不一定从这些领域抽取的大规模无标签公司。为了减轻这一影响,我们提议PERL:一个代表学习模型,扩展基于背景的词嵌入模型,如以枢纽为主的微调的BERT。 PERL超越了22个感官分类领域适应设置的强大基线,改善了内部模型的性能,产生了有效的缩小规模模型,提高了模型稳定性。