The task of automatic text summarization produces a concise and fluent text summary while preserving key information and overall meaning. Recent approaches to document-level summarization have seen significant improvements in recent years by using models based on the Transformer architecture. However, the quadratic memory and time complexities with respect to the sequence length make them very expensive to use, especially with long sequences, as required by document-level summarization. Our work addresses the problem of document-level summarization by studying how efficient Transformer techniques can be used to improve the automatic summarization of very long texts. In particular, we will use the arXiv dataset, consisting of several scientific papers and the corresponding abstracts, as baselines for this work. Then, we propose a novel retrieval-enhanced approach based on the architecture which reduces the cost of generating a summary of the entire document by processing smaller chunks. The results were below the baselines but suggest a more efficient memory a consumption and truthfulness.
翻译:自动文本摘要的任务在保存关键信息和总体含义的同时,产生简明和流畅的文本摘要。最近的文件级摘要办法近年来通过使用基于变换器结构的模型取得了显著的改进。然而,关于序列长度的二次记忆和时间复杂性使其使用非常昂贵,特别是文件级摘要所要求的长序列。我们的工作通过研究如何使用高效的变异器技术改进长文本的自动摘要处理文件级摘要问题。特别是,我们将使用由若干科学论文和相应摘要组成的ArXiv数据集作为这项工作的基线。然后,我们提出一种基于结构的新的检索强化办法,该结构将降低通过处理小块生成整个文件摘要的成本。结果低于基线,但表明可以更有效地存储一个消耗和真实性。