Knowledge Graph Embedding (KGE) is a popular method for KG reasoning and training KGEs with higher dimension are usually preferred since they have better reasoning capability. However, high-dimensional KGEs pose huge challenges to storage and computing resources and are not suitable for resource-limited or time-constrained applications, for which faster and cheaper reasoning is necessary. To address this problem, we propose DualDE, a knowledge distillation method to build low-dimensional student KGE from pre-trained high-dimensional teacher KGE. DualDE considers the dual-influence between the teacher and the student. In DualDE, we propose a soft label evaluation mechanism to adaptively assign different soft label and hard label weights to different triples, and a two-stage distillation approach to improve the student's acceptance of the teacher. Our DualDE is general enough to be applied to various KGEs. Experimental results show that our method can successfully reduce the embedding parameters of a high-dimensional KGE by 7 times - 15 times and increase the inference speed by 2 times - 6 times while retaining a high performance. We also experimentally prove the effectiveness of our soft label evaluation mechanism and two-stage distillation approach via ablation study.
翻译:知识嵌入图(KGE)是KG推理和培训具有较高层面的KGE的流行方法,通常更受欢迎,因为它们具有更好的推理能力。然而,高维KGE对存储和计算资源提出了巨大挑战,不适合资源有限或时间限制的应用,对此,必须更快和廉价的推理。为了解决这一问题,我们提议了DualDE,这是一种知识蒸馏方法,用于从训练有素的高层次教师KGE建立低维学生KGE。DualDE认为教师和学生之间的双重影响。在双维DE中,我们提议了一个软标签评价机制,以适应性地为不同的三重物分配不同的软标签和硬标签重量,并采用两阶段蒸馏方法提高学生对教师的接受程度。我们的双重DE是通用的,足以适用于各种KGEE。实验结果表明,我们的方法可以成功地将高维KGEE的嵌入参数减少7倍至15倍,并将推导速度提高2倍至6倍,同时保留高性的工作表现。我们还实验性地证明了软标签和双级研究的有效性。