While recent work on multilingual language models has demonstrated their capacity for cross-lingual zero-shot transfer on downstream tasks, there is a lack of consensus in the community as to what shared properties between languages enable such transfer. Analyses involving pairs of natural languages are often inconclusive and contradictory since languages simultaneously differ in many linguistic aspects. In this paper, we perform a large-scale empirical study to isolate the effects of various linguistic properties by measuring zero-shot transfer between four diverse natural languages and their counterparts constructed by modifying aspects such as the script, word order, and syntax. Among other things, our experiments show that the absence of sub-word overlap significantly affects zero-shot transfer when languages differ in their word order, and there is a strong correlation between transfer performance and word embedding alignment between languages (e.g., R=0.94 on the task of NLI). Our results call for focus in multilingual models on explicitly improving word embedding alignment between languages rather than relying on its implicit emergence.
翻译:虽然最近关于多语文模式的工作表明,他们有能力在下游任务方面进行跨语言零点转移,但社区内对于不同语言之间的共同属性促成这种转移缺乏共识。涉及两种自然语言的分析往往没有结果,而且自相矛盾,因为语言在许多语言方面同时存在差异。在本文件中,我们进行了大规模的经验研究,通过测量四种不同自然语言与对应语言之间的零点转移,通过修改文字、单词顺序和语法等内容而构建的对应语言之间的零点转移,来孤立各种语言属性的影响。除其他外,我们的实验表明,如果语言在文字顺序上不同,没有分词重叠会严重影响零点转移,而且语言之间的转移性能和语言之间嵌入的文字之间有着密切的关联性(例如,关于NLI的任务的R=0.94)。我们的成果要求以多语种模式为重点,明确改善语言之间的文字融合,而不是依赖其隐含的出现。