Large transformer-based pre-trained language models have achieved impressive performance on a variety of knowledge-intensive tasks and can capture factual knowledge in their parameters. We argue that storing large amounts of knowledge in the model parameters is sub-optimal given the ever-growing amounts of knowledge and resource requirements. We posit that a more efficient alternative is to provide explicit access to contextually relevant structured knowledge to the model and train it to use that knowledge. We present LM-CORE -- a general framework to achieve this -- that allows \textit{decoupling} of the language model training from the external knowledge source and allows the latter to be updated without affecting the already trained model. Experimental results show that LM-CORE, having access to external knowledge, achieves significant and robust outperformance over state-of-the-art knowledge-enhanced language models on knowledge probing tasks; can effectively handle knowledge updates; and performs well on two downstream tasks. We also present a thorough error analysis highlighting the successes and failures of LM-CORE.
翻译:大型变压器的预先培训语言模型在各种知识密集型任务上取得了令人印象深刻的业绩,并能够在其参数中捕捉到事实知识。我们争辩说,鉴于知识和资源需求不断增加,将大量知识储存在模型参数中是不理想的。我们认为,一个更有效的替代办法是向模型提供与背景相关的结构化知识,并培训它使用这种知识。我们提出LM-CORE -- -- 实现这一目标的一般框架 -- -- 允许从外部知识来源进行语言模型培训,使后者得以更新而不影响已经受过培训的模型。实验结果表明,LM-CORE在获得外部知识后,在知识研究任务方面取得了显著和有力的超额业绩,能够有效地处理知识更新,并顺利完成两个下游任务。我们还提出一个彻底的错误分析,突出LM-CORE的成功和失败。