Nowadays, Knowledge graphs (KGs) have been playing a pivotal role in AI-related applications. Despite the large sizes, existing KGs are far from complete and comprehensive. In order to continuously enrich KGs, automatic knowledge construction and update mechanisms are usually utilized, which inevitably bring in plenty of noise. However, most existing knowledge graph embedding (KGE) methods assume that all the triple facts in KGs are correct, and project both entities and relations into a low-dimensional space without considering noise and knowledge conflicts. This will lead to low-quality and unreliable representations of KGs. To this end, in this paper, we propose a general multi-task reinforcement learning framework, which can greatly alleviate the noisy data problem. In our framework, we exploit reinforcement learning for choosing high-quality knowledge triples while filtering out the noisy ones. Also, in order to take full advantage of the correlations among semantically similar relations, the triple selection processes of similar relations are trained in a collective way with multi-task learning. Moreover, we extend popular KGE models TransE, DistMult, ConvE and RotatE with the proposed framework. Finally, the experimental validation shows that our approach is able to enhance existing KGE models and can provide more robust representations of KGs in noisy scenarios.
翻译:目前,知识图(KG)在与AI相关的应用中一直发挥着关键作用。尽管其规模庞大,但现有的KG远非完整和全面。为了不断丰富KG,通常使用自动知识建设和更新机制,这不可避免地带来大量噪音。然而,大多数现有的知识图嵌入方法假定,KG的所有三重事实都是正确的,并且将两个实体和关系投射到一个低维空间,而不考虑噪音和知识冲突。这将导致KG的低质量和不可靠表述。为此,我们在本文件中提出一个通用的多任务强化学习框架,这可以大大缓解繁琐的数据问题。在我们的框架内,我们利用强化学习来选择高质量知识的三倍,同时过滤噪音的三倍。此外,为了充分利用语义上相似关系之间的相互联系,类似关系的三重选择过程将经过集体培训,同时进行多任务学习。此外,我们推广了受欢迎的KGE模型(Trans-E)、DistMult、Conve-E和RotateE),我们可以大大地减轻数据问题。我们利用强化的学习模式来选择高压的模型。最后,可以加强现有的KGGG的模型。