Many AI-related tasks involve the interactions of data in multiple modalities. It has been a new trend to merge multi-modal information into knowledge graph(KG), resulting in multi-modal knowledge graphs (MMKG). However, MMKGs usually suffer from low coverage and incompleteness. To mitigate this problem, a viable approach is to integrate complementary knowledge from other MMKGs. To this end, although existing entity alignment approaches could be adopted, they operate in the Euclidean space, and the resulting Euclidean entity representations can lead to large distortion of KG's hierarchical structure. Besides, the visual information has yet not been well exploited. In response to these issues, in this work, we propose a novel multi-modal entity alignment approach, Hyperbolic multi-modal entity alignment(HMEA), which extends the Euclidean representation to hyperboloid manifold. We first adopt the Hyperbolic Graph Convolutional Networks (HGCNs) to learn structural representations of entities. Regarding the visual information, we generate image embeddings using the densenet model, which are also projected into the hyperbolic space using HGCNs. Finally, we combine the structure and visual representations in the hyperbolic space and use the aggregated embeddings to predict potential alignment results. Extensive experiments and ablation studies demonstrate the effectiveness of our proposed model and its components.
翻译:许多与大赦国际有关的任务涉及多种模式的数据互动,这是将多模式信息合并成知识图表(KG)的新趋势,导致多模式知识图表(MMKG),但是,MMKG通常受到低覆盖率和不完全性的影响。为缓解这一问题,一个可行的办法是将其他MMKG的补充知识结合起来。为此,尽管可以采用现有实体调整方法,但它们在欧克莱底空间运作,由此产生的欧克莱底实体表示可能导致对KG的等级结构的大规模扭曲。此外,视觉信息尚未很好利用。针对这些问题,我们提议采用新的多模式实体调整办法,即超标准多模式多模式实体调整办法(HMEA),将Euclidean代表制扩大到超标准体型体。我们首先采用超标准图表图表网络(HGCN)来学习实体的结构表述。关于视觉信息,我们利用密度网络模型生成图像嵌入CN,该模型也被预测为超标准空间模型和超标准空间模型的模型。最后,我们用超标准模型和超标准空间模型来展示我们的超标准空间模型。