Recently, various methods for representation learning on Knowledge Bases (KBs) have been developed. However, these approaches either only focus on learning the embeddings of the data-level knowledge (ABox) or exhibit inherent limitations when dealing with the concept-level knowledge (TBox), e.g., not properly modelling the structure of the logical knowledge. We present BoxEL, a geometric KB embedding approach that allows for better capturing logical structure expressed in the theories of Description Logic EL++. BoxEL models concepts in a KB as axis-parallel boxes exhibiting the advantage of intersectional closure, entities as points inside boxes, and relations between concepts/entities as affine transformations. We show theoretical guarantees (soundness) of BoxEL for preserving logical structure. Namely, the trained model of BoxEL embedding with loss 0 is a (logical) model of the KB. Experimental results on subsumption reasoning and a real-world application--protein-protein prediction show that BoxEL outperforms traditional knowledge graph embedding methods as well as state-of-the-art EL++ embedding approaches.
翻译:最近,开发了各种知识库(KB)代表学习方法,然而,这些方法要么只是侧重于学习数据级知识(ABox)的嵌入,要么在处理概念级知识(TaTOx)时表现出内在的局限性,例如,没有恰当地模拟逻辑知识的结构。我们介绍了BoxEL,一种几何KB嵌入式方法,可以更好地捕捉逻辑结构,在逻辑 EL++理论中表达。KB中,BoxEL模型概念作为轴-平行箱,展示了交叉封闭的好处,实体作为箱内的点,概念/实体作为近亲转换之间的关系。我们展示了BoxEL的理论保障(稳健性),以维护逻辑结构。也就是说,BoxEL嵌入0号的训练模型是KB的(逻辑)模型。关于子泵推理的实验结果和真实世界应用-蛋蛋预测显示,BoxEL将传统知识图嵌入方法比State-Art EL嵌入式方法还好。