How to effectively adapt neural machine translation (NMT) models according to emerging cases without retraining? Despite the great success of neural machine translation, updating the deployed models online remains a challenge. Existing non-parametric approaches that retrieve similar examples from a database to guide the translation process are promising but are prone to overfit the retrieved examples. However, non-parametric methods are prone to overfit the retrieved examples. In this work, we propose to learn Kernel-Smoothed Translation with Example Retrieval (KSTER), an effective approach to adapt neural machine translation models online. Experiments on domain adaptation and multi-domain machine translation datasets show that even without expensive retraining, KSTER is able to achieve improvement of 1.1 to 1.5 BLEU scores over the best existing online adaptation methods. The code and trained models are released at https://github.com/jiangqn/KSTER.
翻译:如何在不再培训的情况下根据新出现的案例有效地调整神经机翻译模式?尽管神经机翻译取得了巨大成功,但在线更新已部署的模型仍是一个挑战。现有的从数据库中检索类似实例以指导翻译过程的非参数方法很有希望,但很容易超出所检索的实例。然而,非参数方法容易超出所检索的实例。在这项工作中,我们提议学习Kernel-Smoted翻译与实例检索法(KSTER),这是在网上调整神经机翻译模型的有效方法。关于域适应和多多数据机器翻译数据集的实验显示,即使没有昂贵的再培训,KSTER仍然能够改进现有最佳在线适应方法的1.1至1.5 BLEU分数。该代码和经过培训的模式可在https://github.com/jiangqn/KSTER上发布。