Pre-trained models learn contextualized word representations on large-scale text corpus through a self-supervised learning method, which has achieved promising performance after fine-tuning. These models, however, suffer from poor robustness and lack of interpretability. Pre-trained models with knowledge injection, which we call knowledge enhanced pre-trained models (KEPTMs), possess deep understanding and logical reasoning and introduce interpretability to some extent. In this survey, we provide a comprehensive overview of KEPTMs for natural language processing. We first introduce the progress of pre-trained models and knowledge representation learning. Then we systematically categorize existing KEPTMs from three different perspectives. Finally, we outline some potential directions of KEPTMs for future research.
翻译:未经培训的模型通过自我监督的学习方法学习关于大规模文本材料的背景化字表,在微调后取得了有希望的成绩,但这些模型缺乏强健性和可解释性,但是,这些模型缺乏强健性和可解释性,我们称之为知识强化的预培训模型(KEPTMs)的预培训模型具有深刻的理解和逻辑推理,在一定程度上引入了可解释性。在本次调查中,我们全面概述了用于自然语言处理的KEPTMs。我们首先介绍了预先培训的模型和知识代表性学习的进展。然后,我们从三个不同的角度对现有的KEPTMs进行系统分类。最后,我们为今后的研究概述了KEPTMs的一些潜在方向。