Current dialogue summarization systems usually encode the text with a number of general semantic features (e.g., keywords and topics) to gain more powerful dialogue modeling capabilities. However, these features are obtained via open-domain toolkits that are dialog-agnostic or heavily relied on human annotations. In this paper, we show how DialoGPT, a pre-trained model for conversational response generation, can be developed as an unsupervised dialogue annotator, which takes advantage of dialogue background knowledge encoded in DialoGPT. We apply DialoGPT to label three types of features on two dialogue summarization datasets, SAMSum and AMI, and employ pre-trained and non pre-trained models as our summarizes. Experimental results show that our proposed method can obtain remarkable improvements on both datasets and achieves new state-of-the-art performance on the SAMSum dataset.
翻译:目前的对话总和系统通常将文本编码成若干一般语义特征(例如关键词和主题),以获得更强大的对话模型能力;然而,这些特征是通过开放式工具包获得的,这些工具包是对话-不可知的,或严重依赖人的注释。在本文中,我们展示了如何开发DialoGPT,一个经过预先培训的对话响应生成模型,作为不受监督的对话说明器,利用DialoGPT中编码的对话背景知识。我们运用DialoGPT在两个对话综合数据集(SAMSum和AMI)上标出三种特征,并使用预先训练过的和未经训练的模型作为我们的摘要。实验结果表明,我们拟议的方法可以在数据集上取得显著的改进,并在SAMSum数据集上取得新的最新表现。