Despite impressive progress in high-resource settings, Neural Machine Translation (NMT) still struggles in low-resource and out-of-domain scenarios, often failing to match the quality of phrase-based translation. We propose a novel technique that combines back-translation and multilingual NMT to improve performance in these difficult cases. Our technique trains a single model for both directions of a language pair, allowing us to back-translate source or target monolingual data without requiring an auxiliary model. We then continue training on the augmented parallel data, enabling a cycle of improvement for a single model that can incorporate any source, target, or parallel data to improve both translation directions. As a byproduct, these models can reduce training and deployment costs significantly compared to uni-directional models. Extensive experiments show that our technique outperforms standard back-translation in low-resource scenarios, improves quality on cross-domain tasks, and effectively reduces costs across the board.
翻译:尽管在高资源环境下取得了令人瞩目的进展,神经机器翻译(NMT)仍然在低资源和外部情景中挣扎,往往无法达到语句翻译的质量。我们提出了一种新型技术,将反译和多语种NMT结合起来,以提高这些困难案例的绩效。我们的技术为一对语言的双向培训了单一模式,使我们能够在不需要辅助模式的情况下反译源数据或目标单语数据。然后,我们继续就扩大的平行数据进行培训,为能够纳入任何源、目标或平行数据的单一模型提供一个改进周期,以改善两个翻译方向。作为副产品,这些模型可以大大降低培训和部署成本,与单向模式相比。广泛的实验表明,我们的技术在低资源情景中超越标准反译,提高跨领域任务的质量,并有效降低整体成本。