Traditional training paradigms for extractive and abstractive summarization systems always only use token-level or sentence-level training objectives. However, the output summary is always evaluated from summary-level which leads to the inconsistency in training and evaluation. In this paper, we propose a Contrastive Learning based re-ranking framework for one-stage summarization called COLO. By modeling a contrastive objective, we show that the summarization model is able to directly generate summaries according to the summary-level score without additional modules and parameters. Extensive experiments demonstrate that COLO boosts the extractive and abstractive results of one-stage systems on CNN/DailyMail benchmark to 44.58 and 46.33 ROUGE-1 score while preserving the parameter efficiency and inference efficiency. Compared with state-of-the-art multi-stage systems, we save more than 100 GPU training hours and obtaining 3~8 speed-up ratio during inference while maintaining comparable results.
翻译:采掘和抽象总结系统的传统培训模式总是只使用象征性或判决一级的培训目标,然而,产出摘要总是从造成培训和评价不一致的汇总层面来评价,在本文中,我们提议一个阶段总结的对照学习重排框架,称为COLO。通过模拟一个对比性的目标,我们表明汇总模式能够直接根据摘要级评分产生摘要,而没有额外的模块和参数。广泛的实验表明,COLO将CNN/DailyMail基准的一阶段系统的采掘和抽象结果提高到44.58和46.33 ROUGE-1分,同时保持参数效率和推断效率。与最先进的多阶段系统相比,我们节省了100多个GPU培训小时,并在判断期间获得3~8的加速率,同时保持可比较的结果。