Relational knowledge bases (KBs) are established tools for world knowledge representation in machines. While they are advantageous for their precision and interpretability, they usually sacrifice some data modeling flexibility for these advantages because they adhere to a manually engineered schema. In this review, we take a natural language processing perspective to the limitations of KBs, examining how they may be addressed in part by training neural contextual language models (LMs) to internalize and express relational knowledge in free-text form. We propose a novel taxonomy for relational knowledge representation in contextual LMs based on the level of KB supervision provided, considering both works that probe LMs for implicit relational knowledge acquired during self-supervised pretraining on unstructured text alone, and works that explicitly supervise LMs at the level of KB entities and/or relations. We conclude that LMs and KBs are complementary representation tools, as KBs provide a high standard of factual precision which can in turn be flexibly and expressively modeled by LMs, and provide suggestions for future research in this direction.
翻译:关系知识基础(KBs)是世界在机器中体现知识的既定工具,虽然这些基础有利于其精确性和可解释性,但通常会牺牲一些数据模型,为这些优势树立灵活性,因为它们坚持人工设计的计划。在本次审查中,我们从自然语言处理角度来看待KBs的局限性,审查如何通过培训神经背景语言模型(LMs)来部分地解决这些局限性,以便以自由文本的形式内化和表达关系知识。我们提议根据所提供的KB监督水平,为背景LMs中体现关系知识进行新的分类,同时考虑探究LMs仅仅在非结构化文本的自我监督前训练期间获得的隐含关系知识,以及明确监督KB实体和/或关系层面的LMs的工作。我们的结论是,LMs和KBs是辅助性表述工具,因为KBs提供了高标准的事实精确度,反过来可以由LMs以灵活和清晰的方式建模,并为今后朝此方向的研究提供建议。