Based on large-scale pre-trained multilingual representations, recent cross-lingual transfer methods have achieved impressive transfer performances. However, the performance of target languages still lags far behind the source language. In this paper, our analyses indicate such a performance gap is strongly associated with the cross-lingual representation discrepancy. To achieve better cross-lingual transfer performance, we propose the cross-lingual manifold mixup (X-Mixup) method, which adaptively calibrates the representation discrepancy and gives a compromised representation for target languages. Experiments on the XTREME benchmark show X-Mixup achieves 1.8% performance gains on multiple text understanding tasks, compared with strong baselines, and significantly reduces the cross-lingual representation discrepancy.
翻译:在经过培训的大规模多语文代表制的基础上,最近的跨语言转让方法取得了令人印象深刻的转移业绩,然而,目标语言的绩效仍然远远落后于源语言。在本文件中,我们的分析表明,这种绩效差距与跨语言代表制的差异密切相关。为了实现更好的跨语言代表制绩效,我们建议采用跨语言多重混合(X-混合)方法,该方法可适应性地校准代表制差异,并使目标语言的代表性受到损害。对XTREME基准的实验显示,X-Mixup在多文本理解任务上取得了1.8%的绩效收益,而基准则很强,并大大缩小了跨语言代表制的差异。