Many conversation datasets have been constructed in the recent years using crowdsourcing. However, the data collection process can be time consuming and presents many challenges to ensure data quality. Since language generation has improved immensely in recent years with the advancement of pre-trained language models, we investigate how such models can be utilized to generate entire conversations, given only a summary of a conversation as the input. We explore three approaches to generate summary grounded conversations, and evaluate the generated conversations using automatic measures and human judgements. We also show that the accuracy of conversation summarization can be improved by augmenting a conversation summarization dataset with generated conversations.
翻译:近年来,利用众包构建了许多对话数据集,然而,数据收集过程可能耗时,为确保数据质量提出了许多挑战。近年来,随着预先培训的语言模式的进步,语言生成大为改善。我们调查如何利用这些模式生成整个对话,只提供对话摘要作为投入。我们探讨三种方法,以生成基于简要的对话,并利用自动措施和人文判断来评估产生的对话。我们还表明,通过增加对话汇总数据集,可以提高对话总结的准确性。