This paper investigates cross-lingual temporal knowledge graph reasoning problem, which aims to facilitate reasoning on Temporal Knowledge Graphs (TKGs) in low-resource languages by transfering knowledge from TKGs in high-resource ones. The cross-lingual distillation ability across TKGs becomes increasingly crucial, in light of the unsatisfying performance of existing reasoning methods on those severely incomplete TKGs, especially in low-resource languages. However, it poses tremendous challenges in two aspects. First, the cross-lingual alignments, which serve as bridges for knowledge transfer, are usually too scarce to transfer sufficient knowledge between two TKGs. Second, temporal knowledge discrepancy of the aligned entities, especially when alignments are unreliable, can mislead the knowledge distillation process. We correspondingly propose a mutually-paced knowledge distillation model MP-KD, where a teacher network trained on a source TKG can guide the training of a student network on target TKGs with an alignment module. Concretely, to deal with the scarcity issue, MP-KD generates pseudo alignments between TKGs based on the temporal information extracted by our representation module. To maximize the efficacy of knowledge transfer and control the noise caused by the temporal knowledge discrepancy, we enhance MP-KD with a temporal cross-lingual attention mechanism to dynamically estimate the alignment strength. The two procedures are mutually paced along with model training. Extensive experiments on twelve cross-lingual TKG transfer tasks in the EventKG benchmark demonstrate the effectiveness of the proposed MP-KD method.
翻译:本文研究跨语言时间知识图谱推理问题,旨在通过从高资源语言中的时间知识图谱转移知识来促进低资源语言的时间知识图谱推理。在低资源语言中,现有推理方法在严重不完整的时间知识图谱上的表现一直不理想,因此跨语言蒸馏能力变得越来越重要。然而,这涉及两个方面的巨大挑战。首先,作为知识转移桥梁的跨语言对齐通常过于稀缺,无法在两个时间知识图谱之间转移足够的知识。其次,对齐后实体的时间知识差异,特别是当对齐并不可靠时,可能会误导知识蒸馏过程。我们相应地提出了一种互相定步的知识蒸馏模型MP-KD,其中在源时间知识图谱上训练的教师网络可以通过对齐模块引导目标时间知识图谱上的学生网络进行训练。具体而言,为了处理稀缺问题,MP-KD会基于我们的表示模块提取的时间信息生成伪对齐。为了最大程度地提高知识转移效果并控制由时间知识差距导致的噪声,我们使用时间跨语言注意力机制来动态估计对齐强度,两个过程随模型训练而互相定步。通过在 EventKG 基准测试中进行的十二个跨语言时间知识图谱转移任务上的广泛实验,证明了所提出的 MP-KD 方法的有效性。