Task-oriented dialogue systems often employ a Dialogue State Tracker (DST) to successfully complete conversations. Recent state-of-the-art DST implementations rely on schemata of diverse services to improve model robustness and handle zero-shot generalization to new domains [1], however such methods [2, 3] typically require multiple large scale transformer models and long input sequences to perform well. We propose a single multi-task BERT-based model that jointly solves the three DST tasks of intent prediction, requested slot prediction and slot filling. Moreover, we propose an efficient and parsimonious encoding of the dialogue history and service schemata that is shown to further improve performance. Evaluation on the SGD dataset shows that our approach outperforms the baseline SGP-DST by a large margin and performs well compared to the state-of-the-art, while being significantly more computationally efficient. Extensive ablation studies are performed to examine the contributing factors to the success of our model.
翻译:以任务为导向的对话系统往往使用“对话国家跟踪器”来成功完成对话。最近的“数据跟踪”实施依赖于多种服务模式的设计,以提高模型的稳健性,对新领域进行零光化[1],但这类方法[2、3]通常需要多个大型变压器模型和长输入序列才能很好地运行。我们提出了一个单一的多任务BERT模型,共同解决三个“数据跟踪”任务,即意向预测、要求时间档预测和空档填充。此外,我们建议对对话历史和服务系统进行高效和相似的编码,以进一步改进业绩。对“数据采集”数据集的评估表明,我们的方法大大超越了SGP-DST基准,并且与最新技术相比运行良好,同时大大提高了计算效率。我们进行了广泛的实验研究,以研究促进模型成功的因素。