Incorporating prior knowledge can improve existing pre-training models in cloze-style machine reading and has become a new trend in recent studies. Notably, most of the existing models have integrated external knowledge graphs (KG) and transformer-based models, such as BERT into a unified data structure. However, selecting the most relevant ambiguous entities in KG and extracting the best subgraph remains a challenge. In this paper, we propose the LUKE-Graph, a model that builds a heterogeneous graph based on the intuitive relationships between entities in a document without using any external KG. We then use a Relational Graph Attention (RGAT) network to fuse the graph's reasoning information and the contextual representation encoded by the pre-trained LUKE model. In this way, we can take advantage of LUKE, to derive an entity-aware representation; and a graph model - to exploit relation-aware representation. Moreover, we propose Gated-RGAT by augmenting RGAT with a gating mechanism that regulates the question information for the graph convolution operation. This is very similar to human reasoning processing because they always choose the best entity candidate based on the question information. Experimental results demonstrate that the LUKE-Graph achieves state-of-the-art performance on the ReCoRD dataset with commonsense reasoning.
翻译:将先前的知识纳入之前的知识可以改进在凝胶式机器阅读方面的现有培训前模型,并已成为最近研究的新趋势。值得注意的是,大多数现有模型已经将外部知识图(KG)和变压器模型(如BERT)整合到一个统一的数据结构中。然而,选择KG中最相关的模糊实体和提取最佳子集体仍是一个挑战。在本文件中,我们建议使用LUKE-Graph,这是一个基于实体在文件中的直观关系而无需使用外部KG而构建的多元图形模型。然后,我们使用关系图表注意(RGAT)网络将图表的推理学信息和事先经过培训的LUKE模型编码的背景说明整合起来。这样,我们可以利用LUKE来利用最相关的模糊实体,并提取最佳子集体代表;我们建议Gate-RGAT,方法是加强RGAT,建立调控图表卷动操作的问题信息。这与人类的推理学处理非常相似,因为它们总是选择以REDA公司最佳候选数据来展示最佳的测试。</s>