To solve the problem of redundant information and overlapping relations of the entity and relation extraction model, we propose a joint extraction model. This model can directly extract multiple pairs of related entities without generating unrelated redundant information. We also propose a recurrent neural network named Encoder-LSTM that enhances the ability of recurrent units to model sentences. Specifically, the joint model includes three sub-modules: the Named Entity Recognition sub-module consisted of a pre-trained language model and an LSTM decoder layer, the Entity Pair Extraction sub-module which uses Encoder-LSTM network to model the order relationship between related entity pairs, and the Relation Classification sub-module including Attention mechanism. We conducted experiments on the public datasets ADE and CoNLL04 to evaluate the effectiveness of our model. The results show that the proposed model achieves good performance in the task of entity and relation extraction and can greatly reduce the amount of redundant information.
翻译:为了解决实体和关系提取模型的冗余信息和重叠关系问题,我们提议了一个联合提取模型,该模型可以直接提取多个相关实体,而不会产生无关的冗余信息;我们还提议了一个名为Encoder-LSTM的经常性神经网络,以提高经常性单元的示范句号能力;具体地说,该联合模型包括三个分模块:名称实体识别分模块,由预先培训的语言模型和LSTM脱钩层组成;实体Pair提取子模块,利用Encoder-LSTM网络来模拟相关实体对口的排序关系;以及Relation分类子模块,包括注意机制;我们对公共数据集ADE和CONLL04进行了实验,以评价我们模型的有效性;结果显示,拟议的模型在实体任务和关系提取中取得了良好的表现,可以大大减少多余信息的数量。