In simultaneous translation (SimulMT), the most widely used strategy is the wait-k policy thanks to its simplicity and effectiveness in balancing translation quality and latency. However, wait-k suffers from two major limitations: (a) it is a fixed policy that can not adaptively adjust latency given context, and (b) its training is much slower than full-sentence translation. To alleviate these issues, we propose a novel and efficient training scheme for adaptive SimulMT by augmenting the training corpus with adaptive prefix-to-prefix pairs, while the training complexity remains the same as that of training full-sentence translation models. Experiments on two language pairs show that our method outperforms all strong baselines in terms of translation quality and latency.
翻译:在同时翻译(SimulMT)中,最广泛使用的战略是 " 等待-k " 政策,因为它在平衡翻译质量和延缓性方面既简单又有效。然而,等待-k 政策有两大限制:(a) 它是一个固定政策,不能适应性地调整潜伏环境;(b) 其培训比全句翻译要慢得多。为了缓解这些问题,我们建议为适应性模拟MT制定一个新的、高效的培训计划,增加适应性前缀到前缀配对的培训,而培训的复杂性与培训全句翻译模型相同。 对两种语言的实验显示,我们的方法在翻译质量和延缓性方面超过了所有强的基线。