Knowledge-Enhanced Pre-trained Language Models (KEPLMs) are pre-trained models with relation triples injecting from knowledge graphs to improve language understanding abilities. To guarantee effective knowledge injection, previous studies integrate models with knowledge encoders for representing knowledge retrieved from knowledge graphs. The operations for knowledge retrieval and encoding bring significant computational burdens, restricting the usage of such models in real-world applications that require high inference speed. In this paper, we propose a novel KEPLM named DKPLM that Decomposes Knowledge injection process of the Pre-trained Language Models in pre-training, fine-tuning and inference stages, which facilitates the applications of KEPLMs in real-world scenarios. Specifically, we first detect knowledge-aware long-tail entities as the target for knowledge injection, enhancing the KEPLMs' semantic understanding abilities and avoiding injecting redundant information. The embeddings of long-tail entities are replaced by "pseudo token representations" formed by relevant knowledge triples. We further design the relational knowledge decoding task for pre-training to force the models to truly understand the injected knowledge by relation triple reconstruction. Experiments show that our model outperforms other KEPLMs significantly over zero-shot knowledge probing tasks and multiple knowledge-aware language understanding tasks. We further show that DKPLM has a higher inference speed than other competing models due to the decomposing mechanism.
翻译:为保证有效的知识注入,以前的研究将模型与知识编码器结合起来,以代表从知识图形中获取的知识。 知识检索和编码操作带来了巨大的计算负担,限制了这类模型在现实应用中的使用,而这种应用需要高推力速度。 在本文件中,我们提议一个名为DKPLM的新型KEPLM 名为 DKPLM 的新型KEPLM 的嵌入程序,该程序将预先培训的语言模型的知识注入过程从培训前、微调和推断阶段分解为三重,这有利于将KEPLMs应用于现实世界情景中。具体地说,我们首先发现知识识别长尾实体作为知识注入的目标,提高KEPLMs的语义理解能力和避免注入冗余信息。长尾实体的嵌入被相关知识KKP的“假称描述”三重。我们进一步设计相关知识的关联解析任务,用于对部队模型进行预先培训,促进在现实世界情景中应用KEPLMs的应用。我们首先通过三重理解我们的知识,而后展示了其他的零理解,通过三重的模型显示我们的知识,从而展示了其他理解其他知识,从而展示了其他的MS-