Knowledge graph embedding models (KGEMs) are used for various tasks related to knowledge graphs (KGs), including link prediction. They are trained with loss functions that are computed considering a batch of scored triples and their corresponding labels. Traditional approaches consider the label of a triple to be either true or false. However, recent works suggest that all negative triples should not be valued equally. In line with this commonly adopted assumption, we posit that semantically valid negative triples might be high-quality negative triples. As such, loss functions should treat them differently from semantically invalid negative ones. To this aim, we propose semantic-driven versions for the three mostly used loss functions for link prediction. In particular, we treat the scores of negative triples differently by injecting background knowledge about relation domains and ranges into the loss functions. In an extensive and controlled experimental setting, we show that the proposed loss functions systematically provide satisfying results on three public benchmark KGs underpinned with different schemas, which demonstrates both the generality and superiority of our proposed approach. In fact, the proposed loss functions do not only lead to better MRR and Hits@10 values, but also drive KGEMs towards better semantic awareness. This highlights that semantic information globally improves KGEMs, and thus should be incorporated into loss functions whenever such information is available.
翻译:知识嵌入模型(KGEMS)用于与知识图形(KGs)有关的各种任务,包括链接预测,这些模型是知识嵌入模型(KGEMS),它们经过损失功能的培训,在计算损失功能时考虑到一组分数三重和相应的标签。传统方法认为三重标签是真实的或虚假的;然而,最近的工作表明,所有负三重标签不应被同等地估价。根据这一普遍采用的假设,我们认为,具有内在有效性的负三重可能是高质量的负三重;因此,损失功能应不同于词性无效的负三重。因此,我们提议的损失功能应不同于我们拟议方法的一般性和优越性。为此,我们建议三种主要使用的损失函数采用语义驱动的版本来计算链接预测。特别是,我们通过将有关关系领域和范围的背景知识注入损失功能,对负三重的分数进行不同的处理。在广泛和受控制的实验环境中,我们表明,拟议的损失功能系统地提供了三个公共基准的满意的结果,这些基准显示了我们拟议方法的普遍性和优越性。事实上,拟议的损失功能不仅导致更好的MRRR和H@h@hem10的功能,因此,而且还应该改进SGEMES10号。</s>