Incorporating external knowledge into the response generation process is essential to building more helpful and reliable dialog agents. However, collecting knowledge-grounded conversations is often costly, calling for a better pre-trained model for grounded dialog generation that generalizes well w.r.t. different types of knowledge. In this work, we propose KPT (Keyword-guided Pre-Training), a novel self-supervised pre-training method for grounded dialog generation without relying on extra knowledge annotation. Specifically, we use a pre-trained language model to extract the most uncertain tokens in the dialog as keywords. With these keywords, we construct two kinds of knowledge and pre-train a knowledge-grounded response generation model, aiming at handling two different scenarios: (1) the knowledge should be faithfully grounded; (2) it can be selectively used. For the former, the grounding knowledge consists of keywords extracted from the response. For the latter, the grounding knowledge is additionally augmented with keywords extracted from other utterances in the same dialog. Since the knowledge is extracted from the dialog itself, KPT can be easily performed on a large volume and variety of dialogue data. We considered three data sources (open-domain, task-oriented, conversational QA) with a total of 2.5M dialogues. We conduct extensive experiments on various few-shot knowledge-grounded generation tasks, including grounding on dialog acts, knowledge graphs, persona descriptions, and Wikipedia passages. Our comprehensive experiments and analyses demonstrate that KPT consistently outperforms state-of-the-art methods on these tasks with diverse grounding knowledge.
翻译:将外部知识纳入响应生成过程对于建立更有用和更可靠的对话代理器至关重要。 然而,收集基于知识的对话往往成本很高,要求为基础对话生成建立更好的预先培训模式,这种模式能够概括不同的知识类型。在这项工作中,我们建议使用KPT(Keyword-Guided Pre-traction),这是在不依赖额外知识说明的情况下为基础对话生成建立的一种全新的自我监督的训练前方法。具体地,我们使用预先培训的语言模式在对话框中提取最不确定的符号作为关键词。用这些关键字,我们建立两种知识的预培训模式,并预设一个基于知识的全面反应生成模式,目的是处理两种不同的情景:(1)知识应当忠实地基础;(2)可以有选择地使用。对于前者,基础知识由从响应中提取的关键词构成。对于后者,地面知识的添加了从其他语调中提取的关键词。由于从对话中提取的知识本身, KPT可以很容易在大量和多种基于知识的对话框中完成,我们认为,在各种面向的地面对话中,我们用三个数据源来展示了各种数据格式。 我们认为,在地面对话中展示了各种数据格式上展示了各种数据格式。