Cross-lingual adaptation with multilingual pre-trained language models (mPTLMs) mainly consists of two lines of works: zero-shot approach and translation-based approach, which have been studied extensively on the sequence-level tasks. We further verify the efficacy of these cross-lingual adaptation approaches by evaluating their performances on more fine-grained sequence tagging tasks. After re-examining their strengths and drawbacks, we propose a novel framework to consolidate the zero-shot approach and the translation-based approach for better adaptation performance. Instead of simply augmenting the source data with the machine-translated data, we tailor-make a warm-up mechanism to quickly update the mPTLMs with the gradients estimated on a few translated data. Then, the adaptation approach is applied to the refined parameters and the cross-lingual transfer is performed in a warm-start way. The experimental results on nine target languages demonstrate that our method is beneficial to the cross-lingual adaptation of various sequence tagging tasks.
翻译:跨语言适应与多语种预先培训的语言模型(MPTLM)主要包括两行工作:零点办法和基于翻译的方法,已经对顺序层面的任务进行了广泛研究。我们进一步通过评估这些跨语言适应方法在更精细的顺序标记任务方面的绩效来核查这些跨语言适应方法的功效。在重新审查其长处和缺点之后,我们提议了一个新的框架,以整合零点办法和基于翻译的方法,提高适应性能。我们不是简单地用机器翻译的数据来补充源数据,而是用根据一些翻译的数据估计的梯度来调整热度机制,以快速更新 mPTLMs。然后,适应方法应用于精细的参数,跨语言转移以温暖的方式进行。关于9种目标语言的实验结果表明,我们的方法有利于不同顺序标记任务的跨语言适应。