With the recent explosion in popularity of voice assistant devices, there is a growing interest in making them available to user populations in additional countries and languages. However, to provide the highest accuracy and best performance for specific user populations, most existing voice assistant models are developed individually for each region or language, which requires linear investment of effort. In this paper, we propose a general multilingual model framework for Natural Language Understanding (NLU) models, which can help bootstrap new language models faster and reduce the amount of effort required to develop each language separately. We explore how different deep learning architectures affect multilingual NLU model performance. Our experimental results show that these multilingual models can reach same or better performance compared to monolingual models across language-specific test data while require less effort in creating features and model maintenance.
翻译:由于最近语音助理装置的普及程度大增,人们越来越有兴趣向更多国家和语言的用户群体提供这些装置,然而,为了向特定用户群体提供最高准确度和最佳性能,大多数现有的语音助理模型是针对每个区域或语言单独开发的,这需要线性投入努力。在本文件中,我们提议了一个通用的多语种示范框架,用于自然语言理解模型,这样可以更快地吸引新的语言模型,并减少单独开发每种语言所需的努力。我们探讨了不同的深层次学习结构如何影响多语种NLU模型的性能。我们的实验结果表明,这些多语种模型与语言测试数据单一语言模型相比,能够达到相同或更好的性能,同时在创建特征和模型维护方面需要更少的努力。