Opinion summarization is the automatic creation of text reflecting subjective information expressed in multiple documents, such as user reviews of a product. The task is practically important and has attracted a lot of attention. However, due to the high cost of summary production, datasets large enough for training supervised models are lacking. Instead, the task has been traditionally approached with extractive methods that learn to select text fragments in an unsupervised or weakly-supervised way. Recently, it has been shown that abstractive summaries, potentially more fluent and better at reflecting conflicting information, can also be produced in an unsupervised fashion. However, these models, not being exposed to actual summaries, fail to capture their essential properties. In this work, we show that even a handful of summaries is sufficient to bootstrap generation of the summary text with all expected properties, such as writing style, informativeness, fluency, and sentiment preservation. We start by training a conditional Transformer language model to generate a new product review given other available reviews of the product. The model is also conditioned on review properties that are directly related to summaries; the properties are derived from reviews with no manual effort. In the second stage, we fine-tune a plug-in module that learns to predict property values on a handful of summaries. This lets us switch the generator to the summarization mode. We show on Amazon and Yelp datasets that our approach substantially outperforms previous extractive and abstractive methods in automatic and human evaluation.
翻译:意见总和是自动创建反映多种文件中表达的主观信息的文本,例如对产品的用户审查。任务实际上很重要,引起了很多注意。然而,由于摘要制作成本高昂,缺少大量数据,足以培训受监督模型。相反,任务传统上是采用采掘方法,这些方法学会以不受监督或微弱监督的方式选择文本碎片。最近,人们已经表明,抽象摘要,在反映相互矛盾的信息方面可能更加流畅和更好,也可以以一种不受监督的方式产生。然而,这些模型,没有接触到实际摘要,无法捕捉其基本属性。但是,由于摘要制作成本高昂,因此缺少大量的数据集,因此,即使少量摘要也足以吸引生成具有所有预期属性的简要文本,例如写作风格、信息性、流利度和情绪保存。我们从培训一个有条件的变换语言模型开始,以便根据其他可用的产品审查结果进行新的产品审查。这个模型还以审查与摘要直接相关的属性为条件;这些属性来自没有实际摘要的属性,无法捕捉到它们的基本属性。在这项工作中,即使有少量摘要,我们从审查中得出了一些摘要,也可以在预估测的版本中学习了我们之前的模型,然后在模型中学习。