Incorporating prior knowledge into pre-trained language models has proven to be effective for knowledge-driven NLP tasks, such as entity typing and relation extraction. Current pre-training procedures usually inject external knowledge into models by using knowledge masking, knowledge fusion and knowledge replacement. However, factual information contained in the input sentences have not been fully mined, and the external knowledge for injecting have not been strictly checked. As a result, the context information cannot be fully exploited and extra noise will be introduced or the amount of knowledge injected is limited. To address these issues, we propose MLRIP, which modifies the knowledge masking strategies proposed by ERNIE-Baidu, and introduce a two-stage entity replacement strategy. Extensive experiments with comprehensive analyses illustrate the superiority of MLRIP over BERT-based models in military knowledge-driven NLP tasks.
翻译:现有的培训前程序通常通过使用知识掩码、知识融合和知识替换将外部知识注入模型,但投入句中所包含的事实信息尚未充分挖掘,注射的外部知识也未严格检查,因此,不能充分利用背景信息,将引入额外的噪音或注入的知识量有限,为解决这些问题,我们提议MLRIP修改ERNIE-Baidu提出的知识掩码战略,并采用两阶段实体替换战略。全面分析的广泛实验表明,MLRIP在军事知识驱动的NLP任务中优于BERT模型。