In recent years, transformer-based language models have achieved state of the art performance in various NLP benchmarks. These models are able to extract mostly distributional information with some semantics from unstructured text, however it has proven challenging to integrate structured information, such as knowledge graphs into these models. We examine a variety of approaches to integrate structured knowledge into current language models and determine challenges, and possible opportunities to leverage both structured and unstructured information sources. From our survey, we find that there are still opportunities at exploiting adapter-based injections and that it may be possible to further combine various of the explored approaches into one system.
翻译:近年来,以变压器为基础的语言模型在各种国家语言方案基准中达到了最新水平,这些模型能够从非结构化文本中主要从一些语义中提取分布式信息,但事实证明,将结构化信息(如知识图)纳入这些模型具有挑战性。我们研究了将结构化知识纳入当前语言模型的各种办法,确定了挑战,以及利用结构化和非结构化信息来源的可能机会。我们通过调查发现,仍然有机会利用基于适应器的注入,并且有可能进一步将各种探讨过的方法纳入一个系统。