Link prediction based on knowledge graph embedding (KGE) aims to predict new triples to complete knowledge graphs (KGs) automatically. However, recent KGE models tend to improve performance by excessively increasing vector dimensions, which would cause enormous training costs and save storage in practical applications. To address this problem, we first theoretically analyze the capacity of low-dimensional space for KG embeddings based on the principle of minimum entropy. Then, we propose a novel knowledge distillation framework for knowledge graph embedding, utilizing multiple low-dimensional KGE models as teachers. Under a novel iterative distillation strategy, the MulDE model produces soft labels according to training epochs and student performance adaptively. The experimental results show that MulDE can effectively improve the performance and training speed of low-dimensional KGE models. The distilled 32-dimensional models are very competitive compared to some of state-or-the-art (SotA) high-dimensional methods on several commonly-used datasets.
翻译:基于知识图形嵌入(KGE)的链接预测(KGE)旨在预测自动完成知识图形(KGs)的新三重数据。然而,最近的KGE模型往往通过过度增加矢量维度来提高性能,从而产生巨大的培训成本并节省实际应用中的储存量。为了解决这一问题,我们首先从理论上分析基于最小加密原则的KG嵌入的低维空间能力。然后,我们提出了一个知识图形嵌入新知识蒸馏框架,利用多个低维KGE模型作为教师。根据新颖的迭代蒸馏战略,MulDE模型根据培训小区和学生的适应性能生成软标签。实验结果表明,MulDE能够有效地提高低维KGE模型的性能和培训速度。蒸馏的32维模型与一些常用的数据集的州-或艺术(SotA)高维方法相比,具有很强的竞争力。