Existing dialogue modeling methods have achieved promising performance on various dialogue tasks with the aid of Transformer and the large-scale pre-trained language models. However, some recent studies revealed that the context representations produced by these methods suffer the problem of anisotropy. In this paper, we find that the generated representations are also not conversational, losing the conversation structure information during the context modeling stage. To this end, we identify two properties in dialogue modeling, i.e., locality and isotropy, and present a simple method for dialogue representation calibration, namely SimDRC, to build isotropic and conversational feature spaces. Experimental results show that our approach significantly outperforms the current state-of-the-art models on three dialogue tasks across the automatic and human evaluation metrics. More in-depth analyses further confirm the effectiveness of our proposed approach.
翻译:现有对话模式方法在变异器和大规模培训前语言模型的帮助下,在各种对话任务上取得了有希望的成绩,但是,最近的一些研究表明,这些方法产生的背景表现受到厌食症的困扰。在本文中,我们发现,生成的表达方式也不是对话性的,在背景模型阶段失去了对话结构信息。为此,我们确定了对话模式的两个属性,即地点和异位,并提出了对话代表校准的简单方法,即SimDRC,以建立异热带和对话特征空间。实验结果显示,我们的方法大大超越了目前关于自动和人类评价指标的三种对话任务的最新模式。更深入的分析进一步证实了我们拟议方法的有效性。