The adoption of pre-trained language models in task-oriented dialogue systems has resulted in significant enhancements of their text generation abilities. However, these architectures are slow to use because of the large number of trainable parameters and can sometimes fail to generate diverse responses. To address these limitations, we propose two models with auxiliary tasks for response selection - (1) distinguishing distractors from ground truth responses and (2) distinguishing synthetic responses from ground truth labels. They achieve state-of-the-art results on the MultiWOZ 2.1 dataset with combined scores of 107.5 and 108.3 and outperform a baseline with three times more parameters. We publish reproducible code and checkpoints and discuss the effects of applying auxiliary tasks to T5-based architectures.
翻译:在以任务为导向的对话系统中采用预先培训的语言模式,大大提高了其生成文本的能力,然而,由于有大量可训练参数,这些结构的使用速度缓慢,有时无法产生不同的反应,为解决这些局限性,我们提出两个模式,为选择反应作出辅助任务:(1) 区分分散因素和地面真相反应,(2) 区分合成反应和地面真相标签,在多功能组织2.1数据集上取得最新结果,共得分107.5和108.3, 超越基线,比基准参数多三倍。我们出版可复制代码和检查站,讨论对基于T5的结构适用辅助任务的效果。