Nowadays, open-domain dialogue models can generate acceptable responses according to the historical context based on the large-scale pre-trained language models. However, they generally concatenate the dialogue history directly as the model input to predict the response, which we named as the flat pattern and ignores the dynamic information flow across dialogue utterances. In this work, we propose the DialoFlow model, in which we introduce a dynamic flow mechanism to model the context flow, and design three training objectives to capture the information dynamics across dialogue utterances by addressing the semantic influence brought about by each utterance in large-scale pre-training. Experiments on the multi-reference Reddit Dataset and DailyDialog Dataset demonstrate that our DialoFlow significantly outperforms the DialoGPT on the dialogue generation task. Besides, we propose the Flow score, an effective automatic metric for evaluating interactive human-bot conversation quality based on the pre-trained DialoFlow, which presents high chatbot-level correlation ($r=0.9$) with human ratings among 11 chatbots. Code and pre-trained models will be public. \footnote{\url{https://github.com/ictnlp/DialoFlow}}
翻译:目前,开放域对话模式能够根据大规模预先培训语言模式的历史背景产生可接受的反应。然而,这些模式通常将对话历史直接作为预测反应的模型输入,我们称之为平式模式,忽视了对话语句之间的动态信息流动。在这个工作中,我们提出了DialoFlow模式,其中我们引入一个动态流动机制来模拟背景流动,并设计了三个培训目标,以通过解决大规模培训前阶段中每次发言带来的语义影响来捕捉对话语句之间的信息动态。关于多条目 Reddit数据集和DailyDialog数据集的实验表明,我们的 DialoFlow大大地超过了对话生成任务DioloGPT。此外,我们提议了流动评分,这是根据预先培训的DialoFlow软件评估交互式人-bat对话质量的有效自动计分,它展示了11个聊天博特人文评分(r=0.9美元)与11个聊天博特。代码和预先培训模型将公开。