The majority of knowledge graph embedding techniques treat entities and predicates as separate embedding matrices, using aggregation functions to build a representation of the input triple. However, these aggregations are lossy, i.e. they do not capture the semantics of the original triples, such as information contained in the predicates. To combat these shortcomings, current methods learn triple embeddings from scratch without utilizing entity and predicate embeddings from pre-trained models. In this paper, we design a novel fine-tuning approach for learning triple embeddings by creating weak supervision signals from pre-trained knowledge graph embeddings. We develop a method for automatically sampling triples from a knowledge graph and estimating their pairwise similarities from pre-trained embedding models. These pairwise similarity scores are then fed to a Siamese-like neural architecture to fine-tune triple representations. We evaluate the proposed method on two widely studied knowledge graphs and show consistent improvement over other state-of-the-art triple embedding methods on triple classification and triple clustering tasks.
翻译:大部分知识图形嵌入技术将实体和上游作为单独的嵌入矩阵处理,使用聚合功能来构建输入的三重代表。 但是,这些聚合物是亏损的,也就是说,它们不能捕捉原始三重的语义,例如上游所含的信息。为了克服这些缺陷,目前的方法是从零到零学出三重嵌入,而没有利用实体,也没有利用预先培训的模型的上游嵌入。在本文件中,我们设计了一种新的微调方法,通过从预先培训的知识图形嵌入中生成微弱的监督信号来学习三重嵌入。我们开发了一种方法,从知识图表中自动取样三重,并估算其与预先培训的嵌入模型的对等相似之处。这些对称相似的分数随后被输入到一个像暹米色的神经结构中,以微调三重表示。我们评价了两个广泛研究的知识图表的拟议方法,并显示在三重分类和三重组合任务方面与其他最先进的三重嵌入方法相比不断改进。