Aligning sentences in a reference summary with their counterparts in source documents was shown as a useful auxiliary summarization task, notably for generating training data for salience detection. Despite its assessed utility, the alignment step was mostly approached with heuristic unsupervised methods, typically ROUGE-based, and was never independently optimized or evaluated. In this paper, we propose establishing summary-source alignment as an explicit task, while introducing two major novelties: (1) applying it at the more accurate proposition span level, and (2) approaching it as a supervised classification task. To that end, we created a novel training dataset for proposition-level alignment, derived automatically from available summarization evaluation data. In addition, we crowdsourced dev and test datasets, enabling model development and proper evaluation. Utilizing these data, we present a supervised proposition alignment baseline model, showing improved alignment-quality over the unsupervised approach.
翻译:参考摘要中的句子与原始文件中对应的句子对齐被证明是一项有用的辅助性总结任务,主要是为突出的检测生成培训数据。尽管经评估认为该调整步骤很有用,但大多是使用未经监管的超常方法,通常以ROUGE为主,从未独立优化或评估过。在本文件中,我们建议将简要源码调整作为一项明确的任务,同时引入两大新颖之处:(1) 在更准确的标尺范围内应用该模块,(2) 将其作为监督性分类任务处理。为此,我们创建了一套新的提案级调整培训数据集,自动从现有的汇总评价数据中获取。此外,我们还利用了众包式标准格式和测试数据集,促进模型的开发和适当评估。我们利用这些数据,提出了一个监督性提案调整基线模型,显示比未监督的方法更符合一致性质量。