End-to-end relation extraction aims to identify named entities and extract relations between them. Most recent work models these two subtasks jointly, either by casting them in one structured prediction framework, or performing multi-task learning through shared representations. In this work, we present a simple pipelined approach for entity and relation extraction, and establish the new state-of-the-art on standard benchmarks (ACE04, ACE05 and SciERC), obtaining a 1.7%-2.8% absolute improvement in relation F1 over previous joint models with the same pre-trained encoders. Our approach essentially builds on two independent encoders and merely uses the entity model to construct the input for the relation model. Through a series of careful examinations, we validate the importance of learning distinct contextual representations for entities and relations, fusing entity information early in the relation model, and incorporating global context. Finally, we also present an efficient approximation to our approach which requires only one pass of both entity and relation encoders at inference time, achieving an 8-16$\times$ speedup with a slight reduction in accuracy.
翻译:在这项工作中,我们提出了实体和关系提取的简单编程方法,并建立了标准基准的新型技术(ACE04、ACE05和SciERC),在F1方面,与以前与同一培训前编码器的联合模型相比,获得了1.7 %-2.8%的绝对改善。我们的方法基本上以两个独立的编码器为基础,只是利用实体模型来构建关系模型的投入。我们通过一系列仔细的检查,验证了为实体和关系学习不同背景说明的重要性,在关系模型早期使用实体信息,并结合了全球背景。最后,我们还为我们的方法提供了一种有效的近似,在推论期只需要两个实体和关系编码器的一分录,在精确度上略微降低8-16美元。