Zero-shot intent classification is a vital and challenging task in dialogue systems, which aims to deal with numerous fast-emerging unacquainted intents without annotated training data. To obtain more satisfactory performance, the crucial points lie in two aspects: extracting better utterance features and strengthening the model generalization ability. In this paper, we propose a simple yet effective meta-learning paradigm for zero-shot intent classification. To learn better semantic representations for utterances, we introduce a new mixture attention mechanism, which encodes the pertinent word occurrence patterns by leveraging the distributional signature attention and multi-layer perceptron attention simultaneously. To strengthen the transfer ability of the model from seen classes to unseen classes, we reformulate zero-shot intent classification with a meta-learning strategy, which trains the model by simulating multiple zero-shot classification tasks on seen categories, and promotes the model generalization ability with a meta-adapting procedure on mimic unseen categories. Extensive experiments on two real-world dialogue datasets in different languages show that our model outperforms other strong baselines on both standard and generalized zero-shot intent classification tasks.
翻译:零点意向分类是对话系统中一项重要而艰巨的任务,目的是处理大量快速出现的、没有附加说明的培训数据、不熟悉的意图。为了取得更令人满意的业绩,关键要点在于两个方面:提取更好的发音特征,加强模型的概括能力。在本文件中,我们提出一个简单而有效的零点意向分类元学习模式,用于零点意向分类。为更好地学习语句的语义表达,我们引入一个新的混合关注机制,通过同时利用分布信号的注意和多层感官的注意,将相关字发音模式编码起来。为了加强模型从可见的班级向看不见的班级的转移能力,我们用元学习战略重新拟订零点意向分类,通过模拟对已见类别进行多重零点分类任务来培训模型,并推广模型在模拟的看不见类别上采用元适应程序的一般能力。关于两个真实世界对话数据集的不同语言的广泛实验显示,我们的模型在标准和普遍零点意向分类上都比其他强的基线要强。