Intent detection and slot filling are two main tasks for building a spoken language understanding (SLU) system. The two tasks are closely related and the information of one task can be utilized in the other task. Previous studies either model the two tasks separately or only consider the single information flow from intent to slot. None of the prior approaches model the bidirectional connection between the two tasks simultaneously. In this paper, we propose a Co-Interactive Transformer to consider the cross-impact between the two tasks. Instead of adopting the self-attention mechanism in vanilla Transformer, we propose a co-interactive module to consider the cross-impact by building a bidirectional connection between the two related tasks. In addition, the proposed co-interactive module can be stacked to incrementally enhance each other with mutual features. The experimental results on two public datasets (SNIPS and ATIS) show that our model achieves the state-of-the-art performance with considerable improvements (+3.4% and +0.9% on overall acc). Extensive experiments empirically verify that our model successfully captures the mutual interaction knowledge.
翻译:主动探测和空档填充是建立口语理解系统(SLU)的两大主要任务。这两项任务密切相关,一项任务的信息可以用于另一项任务。以前的研究要么将这两项任务分开,要么只考虑从意图到空档的单一信息流动。先前的办法没有一种模式同时将两项任务之间的双向联系作为双向联系的模型。在本文件中,我们提议了一个共同互动变换器,以考虑这两项任务之间的交叉影响。我们提议了一个共同互动模块,而不是在香草变换器中采用自留机制,而是通过在两项相关任务之间建立双向联系来考虑交叉影响。此外,拟议的共同互动模块可以叠叠,以相互增强彼此的特性。两个公共数据集(SNIPS和ATIS)的实验结果显示,我们的模型在取得最新业绩的同时取得了相当大的改进(+3.4%和在总体ACC上+0.9%)。我们进行了广泛的实验,以经验验证我们的模型成功捕捉到相互互动知识。