Reasoning with knowledge expressed in natural language and Knowledge Bases (KBs) is a major challenge for Artificial Intelligence, with applications in machine reading, dialogue, and question answering. General neural architectures that jointly learn representations and transformations of text are very data-inefficient, and it is hard to analyse their reasoning process. These issues are addressed by end-to-end differentiable reasoning systems such as Neural Theorem Provers (NTPs), although they can only be used with small-scale symbolic KBs. In this paper we first propose Greedy NTPs (GNTPs), an extension to NTPs addressing their complexity and scalability limitations, thus making them applicable to real-world datasets. This result is achieved by dynamically constructing the computation graph of NTPs and including only the most promising proof paths during inference, thus obtaining orders of magnitude more efficient models. Then, we propose a novel approach for jointly reasoning over KBs and textual mentions, by embedding logic facts and natural language sentences in a shared embedding space. We show that GNTPs perform on par with NTPs at a fraction of their cost while achieving competitive link prediction results on large datasets, providing explanations for predictions, and inducing interpretable models. Source code, datasets, and supplementary material are available online at https://github.com/uclnlp/gntp.
翻译:以自然语言和知识库(KBs)表达的知识是人工智能的一大挑战,在机器阅读、对话和回答问题方面的应用是人工智能。 联合学习文本表达和转换的一般神经结构非常缺乏数据效率,很难分析其推理过程。这些问题通过诸如神经理论和知识库(NTPs)等端到端的不同推理系统加以解决,尽管它们只能用于小规模的象征性KBs。在本文中,我们首先提议将贪婪NTP(GTPs)扩展至NTP(GNTPs),以解决其复杂性和可缩放性限制,从而使它们适用于现实世界数据集。通过动态地构建NTPs的计算图,只包括推断过程中最有希望的证明路径,从而获得更有效率的模型。然后,我们提出一种新颖的方法,通过将逻辑事实和自然语言句嵌入共同嵌入空间。我们显示,NTPs(GTPs)与NTPs在可竞争性的在线预测结果上进行,同时提供可获取的源代码的大规模数据解释。