The performance of Neural Machine Translation (NMT) systems often suffers in low-resource scenarios where sufficiently large-scale parallel corpora cannot be obtained. Pre-trained word embeddings have proven to be invaluable for improving performance in natural language analysis tasks, which often suffer from paucity of data. However, their utility for NMT has not been extensively explored. In this work, we perform five sets of experiments that analyze when we can expect pre-trained word embeddings to help in NMT tasks. We show that such embeddings can be surprisingly effective in some cases -- providing gains of up to 20 BLEU points in the most favorable setting.
翻译:神经机器翻译系统(NMT)的性能往往在低资源情况下受到影响,因为无法取得足够大规模的平行公司;经过培训的字嵌入证明对于改善自然语言分析任务(通常缺乏数据)的性能非常宝贵;然而,这些系统对NMT的效用尚未广泛探讨;在这项工作中,我们进行了五套实验,分析何时可以预期经过培训的字嵌入有助于NMT的任务;我们表明,在某些情况下,这种嵌入可能出乎意料地有效 -- -- 在最有利的环境下,可达到20个BLEU点。