Recent research in neural machine translation has explored flexible generation orders, as an alternative to left-to-right generation. However, training non-monotonic models brings a new complication: how to search for a good ordering when there is a combinatorial explosion of orderings arriving at the same final result? Also, how do these automatic orderings compare with the actual behaviour of human translators? Current models rely on manually built biases or are left to explore all possibilities on their own. In this paper, we analyze the orderings produced by human post-editors and use them to train an automatic post-editing system. We compare the resulting system with those trained with left-to-right and random post-editing orderings. We observe that humans tend to follow a nearly left-to-right order, but with interesting deviations, such as preferring to start by correcting punctuation or verbs.
翻译:最近对神经机翻译的研究探索了灵活的发电订单,作为左向右转一代的替代。然而,培训非热量模型带来了新的复杂因素:当到达同样最终结果的订单发生组合爆炸时,如何寻找良好的订单?此外,这些自动订单如何与翻译的实际行为相比较?目前的模型依靠人工制造的偏差,或者只能自己探索所有可能性。在本文中,我们分析了人类编辑后产生的订单,并用它们来培训自动编辑后系统。我们比较了由此产生的系统与受过左对右和随机编辑后订单培训的系统。我们观察到,人类往往遵循近乎左对右的顺序,但有有趣的偏差,例如倾向于从纠正标点或动词开始。