Recently, the practical deployment of open-domain dialogue systems has been plagued by the knowledge issue of information deficiency and factual inaccuracy. To this end, we introduce PLATO-K based on two-stage dialogic learning to strengthen internal knowledge memorization and external knowledge exploitation. In the first stage, PLATO-K learns through massive dialogue corpora and memorizes essential knowledge into model parameters. In the second stage, PLATO-K mimics human beings to search for external information and to leverage the knowledge in response generation. Extensive experiments reveal that the knowledge issue is alleviated significantly in PLATO-K with such comprehensive internal and external knowledge enhancement. Compared to the existing state-of-the-art Chinese dialogue model, the overall engagingness of PLATO-K is improved remarkably by 36.2% and 49.2% on chit-chat and knowledge-intensive conversations.
翻译:最近,开放域对话系统的实际部署一直受到信息不足和事实不准确知识问题的困扰。为此,我们引入了基于两阶段对话学习的PLATO-K,以加强内部知识记忆和外部知识开发。在第一阶段,PLATO-K通过大规模对话学习,将基本知识记为模型参数。在第二阶段,PLATO-K模仿人寻找外部信息并利用知识生成反应。广泛的实验显示,知识问题在PLATO-K中大大缓解,并全面增进内部和外部知识。与现有的最先进的中国对话模式相比,PLATO-K的总体参与率显著提高36.2%和49.2%的热聊天和知识密集型对话。