Machine translation systems require semantic knowledge and grammatical understanding. Neural machine translation (NMT) systems often assume this information is captured by an attention mechanism and a decoder that ensures fluency. Recent work has shown that incorporating explicit syntax alleviates the burden of modeling both types of knowledge. However, requiring parses is expensive and does not explore the question of what syntax a model needs during translation. To address both of these issues we introduce a model that simultaneously translates while inducing dependency trees. In this way, we leverage the benefits of structure while investigating what syntax NMT must induce to maximize performance. We show that our dependency trees are 1. language pair dependent and 2. improve translation quality.
翻译:机器翻译系统需要语义学知识和语法理解。神经机器翻译系统常常假定这种信息被一个关注机制和确保流畅的解码器所捕捉。最近的工作表明,采用明确的语法可以减轻两种类型的知识的建模负担。然而,要求剖析费用昂贵,并不探讨翻译时模式需要的语义问题。为了解决这两个问题,我们引入了一种同时翻译并诱发依赖性树的模式。这样,我们在调查什么语法NMT必须诱发最大限度的性能时,就利用结构的好处。我们表明,我们的依赖性树是1对语言,2对语言有依赖性,2提高翻译质量。