Representation learning in the form of semantic embeddings has been successfully applied to a variety of tasks in natural language processing and knowledge graphs. Recently, there has been growing interest in developing similar methods for learning embeddings of entire ontologies. We propose Box$^2$EL, a novel method for representation learning of ontologies in the Description Logic EL++, which represents both concepts and roles as boxes (i.e. axis-aligned hyperrectangles), such that the logical structure of the ontology is preserved. We theoretically prove the soundness of our model and conduct an extensive empirical evaluation, in which we achieve state-of-the-art results in subsumption prediction, link prediction, and deductive reasoning. As part of our evaluation, we introduce a novel benchmark for evaluating EL++ embedding models on predicting subsumptions involving both atomic and complex concepts.
翻译:在自然语言处理和知识图表中,以语义嵌入形式进行的代表学习成功地应用于各种任务。最近,人们越来越有兴趣制定类似的方法来学习整个本体的嵌入。我们提议了Box$2$EL,这是在描述逻辑EL++中代表本体学学习的新颖方法,它既代表了概念,也代表了作为箱的作用(即轴轴对齐的超矩),因此本体学的逻辑结构得以保持。我们理论上证明了我们模型的健全性,并进行了广泛的经验性评估,在这种评估中,我们实现了子假设预测、链接预测和推理方面最先进的结果。作为评估的一部分,我们引入了一个新的基准,用于评价用于预测包含原子和复杂概念的子集的 EL++嵌入模型。