Previous works have shown that contextual information can improve the performance of neural machine translation (NMT). However, most existing document-level NMT methods only consider a few number of previous sentences. How to make use of the whole document as global contexts is still a challenge. To address this issue, we hypothesize that a document can be represented as a graph that connects relevant contexts regardless of their distances. We employ several types of relations, including adjacency, syntactic dependency, lexical consistency, and coreference, to construct the document graph. Then, we incorporate both source and target graphs into the conventional Transformer architecture with graph convolutional networks. Experiments on various NMT benchmarks, including IWSLT English--French, Chinese-English, WMT English--German and Opensubtitle English--Russian, demonstrate that using document graphs can significantly improve the translation quality. Extensive analysis verifies that the document graph is beneficial for capturing discourse phenomena.
翻译:过去的工作表明,背景信息可以改善神经机翻译(NMT)的性能。然而,大多数现有的文件级NMT方法只考虑前几个句子。如何将整个文件作为全球背景加以利用仍是一个挑战。为解决这一问题,我们假设文件可以作为图表来代表,将相关背景联系起来,而不论其距离如何。我们使用几种类型的关系来构建文档图表,包括相近性、综合依赖性、词汇一致性和共同参照。然后,我们将源和目标图表都纳入常规变换器结构中,并配有图形相联网络。关于各种NMT基准的实验,包括IWSLT英语-法语、中文-英语、WMT英语、德语和英语-俄语开放字幕英语-俄语,表明使用文档图表可以大大改进翻译质量。广泛的分析证实文件图有助于捕捉谈话现象。