Knowledge Graph (KG) and attention mechanism have been demonstrated effective in introducing and selecting useful information for weakly supervised methods. However, only qualitative analysis and ablation study are provided as evidence. In this paper, we contribute a dataset and propose a paradigm to quantitatively evaluate the effect of attention and KG on bag-level relation extraction (RE). We find that (1) higher attention accuracy may lead to worse performance as it may harm the model's ability to extract entity mention features; (2) the performance of attention is largely influenced by various noise distribution patterns, which is closely related to real-world datasets; (3) KG-enhanced attention indeed improves RE performance, while not through enhanced attention but by incorporating entity prior; and (4) attention mechanism may exacerbate the issue of insufficient training data. Based on these findings, we show that a straightforward variant of RE model can achieve significant improvements (6% AUC on average) on two real-world datasets as compared with three state-of-the-art baselines. Our codes and datasets are available at https://github.com/zig-kwin-hu/how-KG-ATT-help.
翻译:知识图(KG)和关注机制在引进和选择有用信息以用于监管不力的方法方面证明是有效的,然而,只提供定性分析和模拟研究作为证据,在本文件中,我们提供数据集并提出模式,对关注和KG对包级关系提取(RE)的影响进行定量评估。我们发现:(1) 更高的关注准确性可能导致业绩恶化,因为这可能损害模型提取实体提及特征的能力;(2) 关注的表现在很大程度上受到各种噪音分布模式的影响,这种模式与真实世界数据集密切相关;(3) KG-enhanced 关注确实改善了RE的绩效,但不是通过加强关注,而是通过纳入实体;(4) 关注机制可能会加剧培训数据不足的问题。根据这些调查结果,我们表明,与三个最先进的基线相比,一个直接的RE模型模型模式可以在两个真实世界数据集上实现显著改进(平均6% AUC)。我们的代码和数据集可在https://github.com/zig-kwin-hu/how-KATT-chel)。