Integrating extracted knowledge from the Web to knowledge graphs (KGs) can facilitate tasks like question answering. We study relation integration that aims to align free-text relations in subject-relation-object extractions to relations in a target KG. To address the challenge that free-text relations are ambiguous, previous methods exploit neighbor entities and relations for additional context. However, the predictions are made independently, which can be mutually inconsistent. We propose a two-stage Collective Relation Integration (CoRI) model, where the first stage independently makes candidate predictions, and the second stage employs a collective model that accesses all candidate predictions to make globally coherent predictions. We further improve the collective model with augmented data from the portion of the target KG that is otherwise unused. Experiment results on two datasets show that CoRI can significantly outperform the baselines, improving AUC from .677 to .748 and from .716 to .780, respectively.
翻译:将从网上提取的知识与知识图集(KGs)相结合,可以促进诸如回答问题等任务。我们研究将关系整合,目的是在主题关系-对象提取过程中将自由文本关系与目标KG的关系统一起来。为了应对自由文本关系含混不清的挑战,以往的方法利用邻国实体和关系来创造更多背景。然而,预测是独立的,可以相互不一致。我们提议了一个两阶段集体关系整合模式,第一阶段独立地作出候选人预测,第二阶段采用集体模式,利用所有候选预测进行全球一致预测。我们进一步改进集体模式,从目标KG部分获得更多数据,否则无法使用。两个数据集的实验结果显示,CORI可以大大超过基线,分别从677到748和从716到780改进AUC。