Abstractive summarization models are typically pre-trained on large amounts of generic texts, then fine-tuned on tens or hundreds of thousands of annotated samples. However, in opinion summarization, large annotated datasets of reviews paired with reference summaries are not available and would be expensive to create. This calls for fine-tuning methods robust to overfitting on small datasets. In addition, generically pre-trained models are often not accustomed to the specifics of customer reviews and, after fine-tuning, yield summaries with disfluencies and semantic mistakes. To address these problems, we utilize an efficient few-shot method based on adapters which, as we show, can easily store in-domain knowledge. Instead of fine-tuning the entire model, we add adapters and pre-train them in a task-specific way on a large corpus of unannotated customer reviews, using held-out reviews as pseudo summaries. Then, fine-tune the adapters on the small available human-annotated dataset. We show that this self-supervised adapter pre-training improves summary quality over standard fine-tuning by 2.0 and 1.3 ROUGE-L points on the Amazon and Yelp datasets, respectively. Finally, for summary personalization, we condition on aspect keyword queries, automatically created from generic datasets. In the same vein, we pre-train the adapters in a query-based manner on customer reviews and then fine-tune them on annotated datasets. This results in better-organized summary content reflected in improved coherence and fewer redundancies.
翻译:抽象总和模型通常对大量通用文本进行预先培训,然后对数万或数十万份附加说明的样本进行微调;然而,在意见总结中,大量附带参考摘要的审查附加说明的数据集并不具备,而且要创建成本高昂。这要求采用完善的微调方法,以过度配置小数据集;此外,一般预培训模型往往不适应客户审查的具体内容,在微调后,以不易变和语义错误的形式编写摘要摘要。为了解决这些问题,我们采用基于适应的高效少发方法,根据适应软件,我们使用高效的少发数点方法,如我们所显示的,这些方法很容易存储在主页内知识。我们不是对整个模型进行微调,而是在大量未加注解的客户审查中增加调整和预先培训这些方法。此外,一般调整前审查往往不适应于客户审查的具体内容,然后在基于小的基于人文的调整数据集中对适应者进行微调校准。