Low-resource Multilingual Neural Machine Translation (MNMT) is typically tasked with improving the translation performance on one or more language pairs with the aid of high-resource language pairs. In this paper, we propose two simple search based curricula -- orderings of the multilingual training data -- which help improve translation performance in conjunction with existing techniques such as fine-tuning. Additionally, we attempt to learn a curriculum for MNMT from scratch jointly with the training of the translation system with the aid of contextual multi-arm bandits. We show on the FLORES low-resource translation dataset that these learned curricula can provide better starting points for fine tuning and improve overall performance of the translation system.
翻译:低资源多语言多语种机器翻译(MNMT)通常负责在高资源语言对子的帮助下,改善一种或多种语言对子的翻译绩效。在本文件中,我们提议了两个简单的基于搜索的课程 -- -- 多语种培训数据顺序 -- -- 与微调等现有技术一起帮助提高翻译绩效。此外,我们试图从零开始学习MNMT课程,同时在相关多武器强盗的帮助下对翻译系统进行培训。我们在FLORES的低资源翻译数据集中显示,这些学习的课程可以提供更好的起点,用于微调和改进翻译系统的总体绩效。