Dialogue state tracking (DST) aims to convert the dialogue history into dialogue states which consist of slot-value pairs. As condensed structural information memorizing all history information, the dialogue state in the last turn is typically adopted as the input for predicting the current state by DST models. However, these models tend to keep the predicted slot values unchanged, which is defined as state momentum in this paper. Specifically, the models struggle to update slot values that need to be changed and correct wrongly predicted slot values in the last turn. To this end, we propose MoNET to tackle state momentum via noise-enhanced training. First, the previous state of each turn in the training data is noised via replacing some of its slot values. Then, the noised previous state is used as the input to learn to predict the current state, improving the model's ability to update and correct slot values. Furthermore, a contrastive context matching framework is designed to narrow the representation distance between a state and its corresponding noised variant, which reduces the impact of noised state and makes the model better understand the dialogue history. Experimental results on MultiWOZ datasets show that MoNET outperforms previous DST methods. Ablations and analysis verify the effectiveness of MoNET in alleviating state momentum and improving anti-noise ability.
翻译:对话状态跟踪( DST) 旨在将对话历史转换成由空值对等组成的对话状态。 随着压缩结构信息对全部历史信息进行校正, 对话状态在最后转折时通常被DST模型作为预测当前状态的输入。 然而, 这些模型倾向于保持预测的空档值不变, 也就是本文中的国家动力。 具体地说, 模型努力更新需要修改的空档值, 并在最后转弯中纠正错误预测的空档值。 为此, 我们建议莫内特通过噪声强化培训来应对国家势头。 首先, 培训数据中的每个转折点的先前状态通过替换其一些空档值而被忽略。 然后, 将前转点化状态用作学习预测当前状态的投入, 提高模型更新和纠正空档值的能力。 此外, 对比性环境匹配框架旨在缩小州与相应零度变量之间的代表距离, 从而降低新状态的影响, 并使模型更好地了解对话历史。 在多WOZ 数据库中的实验性结果和 Monet- 变现能力分析显示MINET 之前的变现速度。