Attention-based pre-trained language models such as GPT-2 brought considerable progress to end-to-end dialogue modelling. However, they also present considerable risks for task-oriented dialogue, such as lack of knowledge grounding or diversity. To address these issues, we introduce modified training objectives for language model finetuning, and we employ massive data augmentation via back-translation to increase the diversity of the training data. We further examine the possibilities of combining data from multiples sources to improve performance on the target dataset. We carefully evaluate our contributions with both human and automatic methods. Our model substantially outperforms the baseline on the MultiWOZ data and shows competitive performance with state of the art in both automatic and human evaluation.
翻译:GPT-2等以注意力为基础的预先培训语言模型在端到端对话建模方面取得了相当大的进展,但也给面向任务的对话带来了相当大的风险,例如缺乏基础知识或多样性。为了解决这些问题,我们为语言模型的微调引入了经过修改的培训目标,我们通过回译大量增加数据,以增加培训数据的多样性。我们进一步研究将多种来源的数据结合起来的可能性,以改进目标数据集的绩效。我们仔细评估我们的贡献,同时使用人和自动方法。我们的模型大大超过多功能组织数据的基准,并显示在自动和人评价方面与最新技术相比的竞争性业绩。