Many models learn representations of knowledge graph data by exploiting its low-rank latent structure, encoding known relations between entities and enabling unknown facts to be inferred. To predict whether a relation holds between entities, embeddings are typically compared in the latent space following a relation-specific mapping. Whilst their predictive performance has steadily improved, how such models capture the underlying latent structure of semantic information remains unexplained. Building on recent theoretical understanding of word embeddings, we categorise knowledge graph relations into three types and for each derive explicit requirements of their representations. We show that empirical properties of relation representations and the relative performance of leading knowledge graph representation methods are justified by our analysis.
翻译:许多模型通过利用其低层次潜伏结构、将实体间已知关系编码并能够推断出未知事实来学习知识图形数据的表述。为了预测实体间是否存在某种关系,通常在根据具体关系绘图后在潜藏空间中比较嵌入。虽然它们的预测性能稳步改善,但这类模型如何捕捉语义信息的潜在潜在结构仍然无法解释。根据最近对词嵌入的理论理解,我们将知识图表关系分为三类,对每一种类型都提出了明确要求。我们的分析表明,我们有理由对关系表述的经验性特征和主要知识图表代表方法的相对性能进行分析。