Recent models can generate fluent and grammatical synthetic reviews while accurately predicting user ratings. The generated reviews, expressing users' estimated opinions towards related products, are often viewed as natural language 'rationales' for the jointly predicted rating. However, previous studies found that existing models often generate repetitive, universally applicable, and generic explanations, resulting in uninformative rationales. Further, our analysis shows that previous models' generated content often contain factual hallucinations. These issues call for novel solutions that could generate both informative and factually grounded explanations. Inspired by recent success in using retrieved content in addition to parametric knowledge for generation, we propose to augment the generator with a personalized retriever, where the retriever's output serves as external knowledge for enhancing the generator. Experiments on Yelp, TripAdvisor, and Amazon Movie Reviews dataset show our model could generate explanations that more reliably entail existing reviews, are more diverse, and are rated more informative by human evaluators.
翻译:最近的一些模型可以产生流畅和语法合成的审查,同时准确预测用户的评级。所产生的审查,表达用户对相关产品的估计意见,往往被视为共同预测评级的自然语言“参数”。然而,以往的研究发现,现有模型往往产生重复、普遍适用和通用的解释,导致缺乏信息依据的理由。此外,我们的分析表明,以往模型生成的内容往往包含事实幻觉。这些问题需要新的解决方案,既能产生信息,又能产生事实依据的解释。由于最近成功使用检索到的内容,除了生成的参数知识外,我们建议用个性化检索器来增加生成器,检索器的输出作为加强生成器的外部知识。Yelp、TripAdvisor和亚马孙电影审查数据集的实验显示,我们的模型可以产生解释,更可靠地要求现有的审查,更加多样化,并被人类评价员评为信息更加丰富的信息。