Multi-lingual language models (LM), such as mBERT, XLM-R, mT5, mBART, have been remarkably successful in enabling natural language tasks in low-resource languages through cross-lingual transfer from high-resource ones. In this work, we try to better understand how such models, specifically mT5, transfer *any* linguistic and semantic knowledge across languages, even though no explicit cross-lingual signals are provided during pre-training. Rather, only unannotated texts from each language are presented to the model separately and independently of one another, and the model appears to implicitly learn cross-lingual connections. This raises several questions that motivate our study, such as: Are the cross-lingual connections between every language pair equally strong? What properties of source and target language impact the strength of cross-lingual transfer? Can we quantify the impact of those properties on the cross-lingual transfer? In our investigation, we analyze a pre-trained mT5 to discover the attributes of cross-lingual connections learned by the model. Through a statistical interpretation framework over 90 language pairs across three tasks, we show that transfer performance can be modeled by a few linguistic and data-derived features. These observations enable us to interpret cross-lingual understanding of the mT5 model. Through these observations, one can favorably choose the best source language for a task, and can anticipate its training data demands. A key finding of this work is that similarity of syntax, morphology and phonology are good predictors of cross-lingual transfer, significantly more than just the lexical similarity of languages. For a given language, we are able to predict zero-shot performance, that increases on a logarithmic scale with the number of few-shot target language data points.
翻译:多种语言模式(LMM),如 mBERT, XLM-R, mT5, mT5, mBART, 等多种语言模式(LMM),例如 mBERT, XLM-R, mT5, mT5, mBART, 通过高资源资源的跨语种的跨语言传输, 使低资源语言的自然语言任务变得非常成功。 在这项工作中, 我们试图更好地理解这些模式, 特别是 mT5, 传输语言和语义知识, 尽管在培训前没有提供明确的跨语言的跨语言信号。 相反, 只有每种语言的未加注解文本文本, 而该模式似乎隐含了跨语言的连接。 这提出了几个问题, 激励我们的研究动机是: 每种语言对口语种的跨语言连接是否同样强烈? 源和语义数据流流流流的模型的模型, 能够让我们对一种最精确的排序进行模拟, 通过这些语言和跨数据流流流流的观测, 能够对一种最精确的排序的数据流。