Existing pre-trained language models (PLMs) have demonstrated the effectiveness of self-supervised learning for a broad range of natural language processing (NLP) tasks. However, most of them are not explicitly aware of domain-specific knowledge, which is essential for downstream tasks in many domains, such as tasks in e-commerce scenarios. In this paper, we propose K-PLUG, a knowledge-injected pre-trained language model based on the encoder-decoder transformer that can be transferred to both natural language understanding and generation tasks. We verify our method in a diverse range of e-commerce scenarios that require domain-specific knowledge. Specifically, we propose five knowledge-aware self-supervised pre-training objectives to formulate the learning of domain-specific knowledge, including e-commerce domain-specific knowledge-bases, aspects of product entities, categories of product entities, and unique selling propositions of product entities. K-PLUG achieves new state-of-the-art results on a suite of domain-specific NLP tasks, including product knowledge base completion, abstractive product summarization, and multi-turn dialogue, significantly outperforms baselines across the board, which demonstrates that the proposed method effectively learns a diverse set of domain-specific knowledge for both language understanding and generation tasks.
翻译:在本文件中,我们提议K-PLUG,这是一个以编码器分解器和生成任务为基础的知识输入式预先培训语言模型。我们核实了在一系列需要特定领域知识的电子商务设想方案下方任务中我们采用的方法。具体地说,我们提出了五项了解特定领域的知识的先导性培训目标,以制定具体领域知识的学习方法,包括电子商务特定领域知识库、产品实体的各个方面、产品实体的类别和产品实体独特的销售提议。K-PLUG在一套具体领域、包括产品知识基础完成、抽象产品总结和多功能化对话在内的特定领域任务中取得了新的最新成果,从而大大超越了拟议的具体语言学习基准,从而大大超越了拟议的具体语言基础和不同领域的学习方法。