Pre-trained Language Models (PLMs) which are trained on large text corpus via self-supervised learning method, have yielded promising performance on various tasks in Natural Language Processing (NLP). However, though PLMs with huge parameters can effectively possess rich knowledge learned from massive training text and benefit downstream tasks at the fine-tuning stage, they still have some limitations such as poor reasoning ability due to the lack of external knowledge. Research has been dedicated to incorporating knowledge into PLMs to tackle these issues. In this paper, we present a comprehensive review of Knowledge-Enhanced Pre-trained Language Models (KE-PLMs) to provide a clear insight into this thriving field. We introduce appropriate taxonomies respectively for Natural Language Understanding (NLU) and Natural Language Generation (NLG) to highlight these two main tasks of NLP. For NLU, we divide the types of knowledge into four categories: linguistic knowledge, text knowledge, knowledge graph (KG), and rule knowledge. The KE-PLMs for NLG are categorized into KG-based and retrieval-based methods. Finally, we point out some promising future directions of KE-PLMs.
翻译:通过自我监督的学习方法接受过大量文字材料培训的预先培训的语言模型(PLM),在自然语言处理(NLPM)的各种任务中取得了有希望的成绩,然而,尽管具有巨大参数的PLM能够有效地掌握从大规模培训文本中学到的丰富知识,并在微调阶段使下游任务受益,但是由于缺乏外部知识,这些语言模型仍然有一些局限性,例如推理能力差。研究致力于将知识纳入PLM以解决这些问题。在本文件中,我们全面审查了知识强化的预先培训语言模型(KE-PLMS),以明确了解这个繁荣的领域。我们分别为自然语言理解和自然语言生成(NLG)引进了适当的分类,以突出NLPP的这两项主要任务。关于NLU,我们将知识类型分为四类:语言知识、文字知识、知识图表(KG)和规则知识知识知识。NLG的KE-PLMS(KM)被归类为基于KG和检索方法。最后,我们指出KEPLMS-PLM的一些有希望的未来方向。