Despite the rapid progress of open-domain generation-based conversational agents, most deployed systems treat dialogue contexts as single-turns, while systems dealing with multi-turn contexts are less studied. There is a lack of a reliable metric for evaluating multi-turn modelling, as well as an effective solution for improving it. In this paper, we focus on an essential component of multi-turn generation-based conversational agents: context attention distribution, i.e. how systems distribute their attention on dialogue's context. For evaluation of this component, We introduce a novel attention-mechanism-based metric: DAS ratio. To improve performance on this component, we propose an optimization strategy that employs self-contained distractions. Our experiments on the Ubuntu chatlogs dataset show that models with comparable perplexity can be distinguished by their ability on context attention distribution. Our proposed optimization strategy improves both non-hierarchical and hierarchical models on the proposed metric by about 10% from baselines.
翻译:尽管以开放域为主的一代人对话媒介取得了快速进展,但大多数部署的系统将对话环境视为单一回合,而处理多回合环境的系统则较少研究。缺乏评估多回合建模的可靠衡量标准,以及改进这种建模的有效解决办法。在本文中,我们把重点放在多回合对话媒介的基本组成部分上:背景关注分布,即系统如何在对话背景上分配注意力。为评价这一组成部分,我们采用了一种新的关注-机制衡量标准:DAS比率。为了改进这一组成部分的绩效,我们建议采用一个使用自成一体的分心的优化战略。我们在Ubuntu聊天室数据集的实验表明,对相近的曲解模式,可以用其在背景关注分布上的能力加以区分。我们提议的优化战略将拟议指标的非等级和等级模式从基线上提高大约10%。