Relation Extraction (RE) has attracted increasing attention, but current RE evaluation is limited to in-domain evaluation setups. Little is known on how well a RE system fares in challenging, but realistic out-of-distribution evaluation setups. To address this gap, we propose CrossRE, a new, freely-available cross-domain benchmark for RE, which comprises six distinct text domains and includes multi-label annotations. An additional innovation is that we release meta-data collected during annotation, to include explanations and flags of difficult instances. We provide an empirical evaluation with a state-of-the-art model for relation classification. As the meta-data enables us to shed new light on the state-of-the-art model, we provide a comprehensive analysis on the impact of difficult cases and find correlations between model and human annotations. Overall, our empirical investigation highlights the difficulty of cross-domain RE. We release our dataset, to spur more research in this direction.
翻译:关系提取(RE)已引起越来越多的注意,但目前的RE评价仅限于内部评价设置。对于RE系统在挑战性但现实的分布性外评价设置方面效果如何,我们所知甚少。为了弥补这一差距,我们提议CrossRE, 一个新的、可自由获取的RE跨域基准,由六个不同的文本领域组成,包括多标签说明。另一个创新是,我们发布在批注中收集的元数据,以包括困难事例的解释和标志。我们提供了一个经验性评价,并有一个最先进的关系分类模式。由于元数据使我们能够对最新模式提供新的了解,我们提供了对困难案例影响的全面分析,并找到了模型与人文说明之间的相互关系。总体而言,我们的经验性调查突出了交叉域域的难度。我们发布数据集,以刺激这方面的更多研究。