Multilingual language models have shown decent performance in multilingual and cross-lingual natural language understanding tasks. However, the power of these multilingual models in code-switching tasks has not been fully explored. In this paper, we study the effectiveness of multilingual language models to understand their capability and adaptability to the mixed-language setting by considering the inference speed, performance, and number of parameters to measure their practicality. We conduct experiments in three language pairs on named entity recognition and part-of-speech tagging and compare them with existing methods, such as using bilingual embeddings and multilingual meta-embeddings. Our findings suggest that pre-trained multilingual models do not necessarily guarantee high-quality representations on code-switching, while using meta-embeddings achieves similar results with significantly fewer parameters.
翻译:多语种模式在多语种和跨语种的自然语言理解任务方面表现良好,但是,这些多语种模式在代码转换任务中的力量尚未得到充分探讨,在本文件中,我们研究多语种模式在理解其能力和适应混合语言环境方面的效力,方法是考虑推论速度、性能和衡量其实用性参数的数量。我们用三对语言进行实验,研究名称实体的识别和部分语音标记,并将其与现有方法进行比较,例如使用双语嵌入和多语种元编组。我们的调查结果表明,预先培训的多语种模式不一定保证在代码转换方面高质量的代表性,而使用元编组的参数则要少得多。