State-of-the-art abstractive summarization models generally rely on extensive labeled data, which lowers their generalization ability on domains where such data are not available. In this paper, we present a study of domain adaptation for the abstractive summarization task across six diverse target domains in a low-resource setting. Specifically, we investigate the second phase of pre-training on large-scale generative models under three different settings: 1) source domain pre-training; 2) domain-adaptive pre-training; and 3) task-adaptive pre-training. Experiments show that the effectiveness of pre-training is correlated with the similarity between the pre-training data and the target domain task. Moreover, we find that continuing pre-training could lead to the pre-trained model's catastrophic forgetting, and a learning method with less forgetting can alleviate this issue. Furthermore, results illustrate that a huge gap still exists between the low-resource and high-resource settings, which highlights the need for more advanced domain adaptation methods for the abstractive summarization task.
翻译:最先进的抽象总结模型一般依赖广泛的标签数据,这降低了其在无法获得这些数据的领域的一般能力。在本文中,我们提出对在资源低的情况下对六个不同目标领域的抽象总结任务进行领域调整的研究。具体地说,我们调查了在三种不同情况下大规模基因化模型的第二阶段预培训:(1) 源域前培训;(2) 域适应前培训;和(3) 任务适应前培训。实验表明,培训前培训的有效性与培训前数据和目标领域任务之间的相似性相关。此外,我们发现,继续培训前培训可能导致培训前模式灾难性的遗忘,而不那么遗忘的学习方法可以缓解这一问题。此外,结果显示,在低资源与高资源环境之间仍然存在巨大差距,这突出表明需要为抽象总结任务采用更先进的域适应方法。