Stepping from sentence-level to document-level relation extraction, the research community confronts increasing text length and more complicated entity interactions. Consequently, it is more challenging to encode the key sources of information--relevant contexts and entity types. However, existing methods only implicitly learn to model these critical information sources while being trained for relation extraction. As a result, they suffer the problems of ineffective supervision and uninterpretable model predictions. In contrast, we propose to explicitly teach the model to capture relevant contexts and entity types by supervising and augmenting intermediate steps (SAIS) for relation extraction. Based on a broad spectrum of carefully designed tasks, our proposed SAIS method not only extracts relations of better quality due to more effective supervision, but also retrieves the corresponding supporting evidence more accurately so as to enhance interpretability. By assessing model uncertainty, SAIS further boosts the performance via evidence-based data augmentation and ensemble inference while reducing the computational cost. Eventually, SAIS delivers state-of-the-art relation extraction results on three benchmarks (DocRED, CDR, and GDA) and achieves 5.04% relative gains in F1 score compared to the runner-up in evidence retrieval on DocRED.
翻译:研究界从句级到文件级关系提取,面临越来越多的文字长度和更为复杂的实体互动问题,因此,将信息相关背景和实体类型的关键来源编码起来更具挑战性,然而,现有方法仅含蓄地学习模拟这些关键信息来源,同时接受有关提取的培训;因此,它们遇到监管不力和无法解释模型预测等问题;相反,我们提议明确教授模型,通过监督和加强关系提取的中间步骤(SAIS)来捕捉相关背景和实体类型;根据一系列精心设计的任务,我们拟议的SAIS方法不仅提取出由于更有效的监督而具有更高质量的关系,而且还检索相应的辅助证据,以便提高可解释性;通过评估模型不确定性,SAIS通过基于证据的数据增强和共性推断,进一步提高绩效,同时降低计算成本;最后,SAIS在三个基准(DocRED、CD和GDA)上提供最新的相关提取结果;在F1评分中实现5.04 %的相对RED,相对于后级证据在F1评分中获得5.4%的成绩。