Neural models of dialog rely on generalized latent representations of language. This paper introduces a novel training procedure which explicitly learns multiple representations of language at several levels of granularity. The multi-granularity training algorithm modifies the mechanism by which negative candidate responses are sampled in order to control the granularity of learned latent representations. Strong performance gains are observed on the next utterance retrieval task using both the MultiWOZ dataset and the Ubuntu dialog corpus. Analysis significantly demonstrates that multiple granularities of representation are being learned, and that multi-granularity training facilitates better transfer to downstream tasks.
翻译:对话的神经模式依赖于通用的潜在语言表达方式。本文件引入了一种新的培训程序,明确学习了几个颗粒层次上多种语言的表述方式。多腺培训算法改变了对候选人负面反应进行抽样检查的机制,以控制所学潜在表现方式的颗粒性。在下一个演讲检索任务中,使用多木卫一数据集和Ubuntu对话文集观察到了强大的绩效收益。分析显著地表明,正在学习多种代表方式,而且多腺培训有助于更好地向下游任务转移。