The DocRED dataset is one of the most popular and widely used benchmarks for document-level relation extraction (RE). It adopts a recommend-revise annotation scheme so as to have a large-scale annotated dataset. However, we find that the annotation of DocRED is incomplete, i.e., false negative samples are prevalent. We analyze the causes and effects of the overwhelming false negative problem in the DocRED dataset. To address the shortcoming, we re-annotate 4,053 documents in the DocRED dataset by adding the missed relation triples back to the original DocRED. We name our revised DocRED dataset Re-DocRED. We conduct extensive experiments with state-of-the-art neural models on both datasets, and the experimental results show that the models trained and evaluated on our Re-DocRED achieve performance improvements of around 13 F1 points. Moreover, we conduct a comprehensive analysis to identify the potential areas for further improvement. Our dataset is publicly available at https://github.com/tonytan48/Re-DocRED.
翻译:DocRED数据集是文件级关系提取(RE)中最受欢迎和最广泛使用的基准之一。它采用建议性复读注释办法,以便有一个大规模附加说明的数据集。然而,我们发现DocRED的注解不完整,即虚伪的样本很普遍。我们分析了DocRED数据集中压倒性的虚假负面问题的原因和影响。为了解决缺陷,我们重新在DocRED数据集中重新公布4 053份文件,将错失的关联数增加到原来的DocRED中。我们用我们的名字命名了我们订正的DocRED数据集重新-DocRED。我们对两个数据集都进行了最新的神经模型进行广泛的实验,实验结果显示,对我们的Re-DocRED进行训练和评价的模型在13个F1点上实现了性能改进。此外,我们进行了全面分析,以确定可能进一步改进的领域。我们的数据集可在https://github.com/tonytan48/Re-DocRED上公开查阅。