Intent discovery is a fundamental task in NLP, and it is increasingly relevant for a variety of industrial applications (Quarteroni 2018). The main challenge resides in the need to identify from input utterances novel unseen in-tents. Herein, we propose Z-BERT-A, a two-stage method for intent discovery relying on a Transformer architecture (Vaswani et al. 2017; Devlin et al. 2018), fine-tuned with Adapters (Pfeiffer et al. 2020), initially trained for Natural Language Inference (NLI), and later applied for unknown in-tent classification in a zero-shot setting. In our evaluation, we firstly analyze the quality of the model after adaptive fine-tuning on known classes. Secondly, we evaluate its performance casting intent classification as an NLI task. Lastly, we test the zero-shot performance of the model on unseen classes, showing how Z-BERT-A can effectively perform in-tent discovery by generating intents that are semantically similar, if not equal, to the ground truth ones. Our experiments show how Z-BERT-A is outperforming a wide variety of baselines in two zero-shot settings: known intents classification and unseen intent discovery. The proposed pipeline holds the potential to be widely applied in a variety of application for customer care. It enables automated dynamic triage using a lightweight model that, unlike large language models, can be easily deployed and scaled in a wide variety of business scenarios. Especially when considering a setting with limited hardware availability and performance whereon-premise or low resource cloud deployments are imperative. Z-BERT-A, predicting novel intents from a single utterance, represents an innovative approach for intent discovery, enabling online generation of novel intents. The pipeline is available as an installable python package at the following link: https://github.com/GT4SD/zberta.
翻译:内在发现是NLP的一项根本任务,对于各种工业应用(2018年季度)来说,它越来越重要。主要的挑战在于需要从输入表达式中识别出新颖的不可见内容。在这里,我们提出Z-BERT-A,这是利用变压器结构进行意向发现的一个两阶段方法(Vaswani等人,2017年;Devlin等人,2018年),与调试器(Pfeiffer等人,2020年)进行微调,最初是针对自然语言低度推断(NLI,2020年),后来在零点设置中用于未知的云量分类。在我们评估中,我们首先分析模型的质量,在对已知类进行适应性微调后,我们评价其性能分类为NLILI任务。最后,我们测试该模型在隐形类上的零点性能表现,显示Z-BERT-A如何通过生成简单(如果不相等的话)的立调方法来有效进行动态发现。我们的实验显示,在轨迹-BERT-A中,在深度部署中,将一个已知的深度定位定位的大规模应用一个潜在数据序列显示一个潜在数据序列显示一个潜在的数据转换。