Identifying user intents from natural language utterances is a crucial step in conversational systems that has been extensively studied as a supervised classification problem. However, in practice, new intents emerge after deploying an intent detection model. Thus, these models should seamlessly adapt and classify utterances with both seen and unseen intents -- unseen intents emerge after deployment and they do not have training data. The few existing models that target this setting rely heavily on the scarcely available training data and overfit to seen intents data, resulting in a bias to misclassify utterances with unseen intents into seen ones. We propose RIDE: an intent detection model that leverages commonsense knowledge in an unsupervised fashion to overcome the issue of training data scarcity. RIDE computes robust and generalizable relationship meta-features that capture deep semantic relationships between utterances and intent labels; these features are computed by considering how the concepts in an utterance are linked to those in an intent label via commonsense knowledge. Our extensive experimental analysis on three widely-used intent detection benchmarks shows that relationship meta-features significantly increase the accuracy of detecting both seen and unseen intents and that RIDE outperforms the state-of-the-art model for unseen intents.
翻译:从自然语言语句中识别用户意图是对话系统的关键步骤,已作为监督分类问题进行了广泛研究。然而,在实践中,在部署意图检测模型后出现了新的意图。因此,这些模型应无缝地调整和分类带有可见和看不见意图的言论 -- -- 部署后出现未见意图,它们没有培训数据。针对这一设置的少数现有模型严重依赖稀少的培训数据,并过于适合显示意图数据,从而导致偏差地将隐蔽意图的言论错误划为可观察数据。我们提议,REIDE:一种意图检测模型,以不受监督的方式利用普通知识克服培训数据稀缺问题。REIDE将强健和可概括的关系元性进行编译,以掌握言语与意图标签之间的深层语义关系;这些特征是通过考虑从通识中概念如何与意图标签中的概念相联系,从而计算出来的。我们对三种广泛使用的意向检测基准进行的广泛实验分析表明,关系元性模型大大提高了探测所见和不可见的意向模型的准确性。