Zero-shot translation is a promising direction for building a comprehensive multilingual neural machine translation (MNMT) system. However, its quality is still not satisfactory due to off-target issues. In this paper, we aim to understand and alleviate the off-target issues from the perspective of uncertainty in zero-shot translation. By carefully examining the translation output and model confidence, we identify two uncertainties that are responsible for the off-target issues, namely, extrinsic data uncertainty and intrinsic model uncertainty. Based on the observations, we propose two light-weight and complementary approaches to denoise the training data for model training, and mask out the vocabulary of the off-target languages in inference. Extensive experiments on both balanced and unbalanced datasets show that our approaches significantly improve the performance of zero-shot translation over strong MNMT baselines. Qualitative analyses provide insights into where our approaches reduce off-target translations
翻译:零点翻译是建立综合性多语种神经机翻译(MNMT)系统的一个很有希望的方向,然而,由于目标外的问题,其质量仍然不能令人满意。在本文件中,我们的目标是从零点翻译的不确定性的角度理解和缓解目标外的问题。通过仔细审查翻译产出和模型信心,我们找出了造成目标外问题的两个不确定因素,即外部数据不确定性和内在模型不确定性。根据观察,我们建议了两种轻量和互补的办法,将培训数据用于模型培训,并掩盖了非目标语言的词汇。关于平衡和不平衡数据集的广泛实验表明,我们的办法大大改善了在多国MNMT的强强基线上零点翻译的绩效。定性分析提供了我们减少非目标翻译的方法的洞察力。