Humans usually have conversations by making use of prior knowledge about a topic and background information of the people whom they are talking to. However, existing conversational agents and datasets do not consider such comprehensive information, and thus they have a limitation in generating the utterances where the knowledge and persona are fused properly. To address this issue, we introduce a call For Customized conversation (FoCus) dataset where the customized answers are built with the user's persona and Wikipedia knowledge. To evaluate the abilities to make informative and customized utterances of pre-trained language models, we utilize BART and GPT-2 as well as transformer-based models. We assess their generation abilities with automatic scores and conduct human evaluations for qualitative results. We examine whether the model reflects adequate persona and knowledge with our proposed two sub-tasks, persona grounding (PG) and knowledge grounding (KG). Moreover, we show that the utterances of our data are constructed with the proper knowledge and persona through grounding quality assessment.
翻译:人类通常通过使用他们所交谈的人的话题和背景资料的事先知识进行交谈。 但是,现有的谈话代理人和数据集并不考虑这种全面信息,因此,他们在生成知识和人文融合得当的语句方面受到限制。为了解决这一问题,我们引入了“定制对话(FoCus)”数据集的呼叫,即与用户个人和维基百科知识建立定制的答案。为了评估对培训前语言模型进行信息化和定制表达的能力,我们使用BART和GPT-2以及基于变异器的模型。我们用自动评分来评估其生成能力,并对质量结果进行人文评估。我们检查模型是否反映适当的人文和知识与我们拟议的两个子任务,即人文地面(PG)和知识地面(KG)。此外,我们展示了我们数据的表达方式是用适当的知识和人文通过基础质量评估来构建的。