Recent work pre-training Transformers with self-supervised objectives on large text corpora has shown great success when fine-tuned on downstream NLP tasks including text summarization. However, pre-training objectives tailored for abstractive text summarization have not been explored. Furthermore there is a lack of systematic evaluation across diverse domains. In this work, we propose pre-training large Transformer-based encoder-decoder models on massive text corpora with a new self-supervised objective. In PEGASUS, important sentences are removed/masked from an input document and are generated together as one output sequence from the remaining sentences, similar to an extractive summary. We evaluated our best PEGASUS model on 12 downstream summarization tasks spanning news, science, stories, instructions, emails, patents, and legislative bills. Experiments demonstrate it achieves state-of-the-art performance on all 12 downstream datasets measured by ROUGE scores. Our model also shows surprising performance on low-resource summarization, surpassing previous state-of-the-art results on 6 datasets with only 1000 examples. Finally we validated our results using human evaluation and show that our model summaries achieve human performance on multiple datasets.
翻译:最近对具有大型文本公司自我监督目标的变革者进行了培训前工作前工作,在对下游NLP任务(包括文本摘要)进行微调时,培训前工作取得了巨大成功;然而,尚未探讨为抽象文本摘要而专门设计的训练前目标;此外,还缺乏对不同领域的系统评价;在这项工作中,我们提议对大型文本公司进行大规模文本公司大型变压器基于编码器脱钩模型的培训前工作前工作,并采用新的自我监督目标;在PEGASUS中,从投入文件中删除/装上重要句子,并将其作为剩余句子(类似于抽取摘要)的一个产出序列一起产生;我们评估了我们12个下游总结任务的最佳PEGASUS模式,涵盖新闻、科学、故事、指示、电子邮件、专利和立法法案。实验表明,它在所有12个下游数据集上都取得了最新业绩,由ROUGE分测得的分数。我们的模型还显示,在低资源总和6个模型上超过以往的状态结果,与提取的概要相类似。我们最后用1,000个实例来验证我们人类的成绩。