The non-autoregressive models have boosted the efficiency of neural machine translation through parallelized decoding at the cost of effectiveness when comparing with the autoregressive counterparts. In this paper, we claim that the syntactic and semantic structures among natural language are critical for non-autoregressive machine translation and can further improve the performance. However, these structures are rarely considered in the existing non-autoregressive models. Inspired by this intuition, we propose to incorporate the explicit syntactic and semantic structures of languages into a non-autoregressive Transformer, for the task of neural machine translation. Moreover, we also consider the intermediate latent alignment within target sentences to better learn the long-term token dependencies. Experimental results on two real-world datasets (i.e., WMT14 En-De and WMT16 En-Ro) show that our model achieves a significantly faster speed, as well as keeps the translation quality when compared with several state-of-the-art non-autoregressive models.
翻译:与自动递减的对等方相比,非自动递减模型通过平行解码提高了神经机翻译的效率,在与自动递减的对等方比较时以效率成本计算,提高了神经机翻译的效率。在本文中,我们声称自然语言的合成和语义结构对于非自动机器翻译至关重要,可以进一步改进性能。然而,现有的非非自动递减模型很少考虑这些结构。受这一直觉的启发,我们提议将语言的明确的合成和语义结构纳入非自动递减变异器,用于神经机翻译任务。此外,我们还考虑在目标句内进行中间潜值调整,以更好地学习长期象征性依赖性。两个真实世界数据集(即WMT14 En-De和WMT16 En-Ro)的实验结果显示,我们的模型实现的速度要快得多,并且与几个最先进的非侵略模型相比,还保持翻译质量。