A dialogue is essentially a multi-turn interaction among interlocutors. Effective evaluation metrics should reflect the dynamics of such interaction. Existing automatic metrics are focused very much on the turn-level quality, while ignoring such dynamics. To this end, we propose DynaEval, a unified automatic evaluation framework which is not only capable of performing turn-level evaluation, but also holistically considers the quality of the entire dialogue. In DynaEval, the graph convolutional network (GCN) is adopted to model a dialogue in totality, where the graph nodes denote each individual utterance and the edges represent the dependency between pairs of utterances. A contrastive loss is then applied to distinguish well-formed dialogues from carefully constructed negative samples. Experiments show that DynaEval significantly outperforms the state-of-the-art dialogue coherence model, and correlates strongly with human judgements across multiple dialogue evaluation aspects at both turn and dialogue level.
翻译:有效的评价指标应该反映这种互动的动态。现有的自动衡量指标非常侧重于转折点质量,而忽略了这种动态。为此,我们提议DynaEval,这是一个统一的自动评价框架,不仅能够进行转折点评价,而且能够通盘考虑整个对话的质量。在DynaEval,图形共变网络(GCN)被采用来模拟全面的对话,其中图形节点代表了每个个人的话语和边缘代表了两种话语之间的依赖性。随后将对比性损失用于区分完善的对话和精心构建的负面样本。实验显示,DynaEval大大超越了最先进的对话一致性模式,与人对交替和对话层面多个对话评价方面的判断密切相关。