Neural Machine Translation systems built on top of Transformer-based architectures are routinely improving the state-of-the-art in translation quality according to word-overlap metrics. However, a growing number of studies also highlight the inherent gender bias that these models incorporate during training, which reflects poorly in their translations. In this work, we investigate whether these models can be instructed to fix their bias during inference using targeted, guided instructions as contexts. By translating relevant contextual sentences during inference along with the input, we observe large improvements in reducing the gender bias in translations, across three popular test suites (WinoMT, BUG, SimpleGen). We further propose a novel metric to assess several large pre-trained models (OPUS-MT, M2M-100) on their sensitivity towards using contexts during translation to correct their biases. Our approach requires no fine-tuning and thus can be used easily in production systems to de-bias translations from stereotypical gender-occupation bias 1. We hope our method, along with our metric, can be used to build better, bias-free translation systems.
翻译:在以变换器为基础的结构之上所建的神经机器翻译系统,根据多字标准,经常改善翻译质量的先进水平;然而,越来越多的研究还突出这些模型在培训过程中所固有的性别偏见,这在翻译中表现得很差;在这项工作中,我们调查这些模型是否可以指示在推断过程中使用有针对性的、指导性的指示,纠正其偏见;在推断过程中,通过翻译相关的背景句子以及投入,我们观察到在减少翻译中的性别偏见方面大有改进,这体现在三个流行的测试套房(WinoMT、BUG、SimpleGen)中;我们进一步提出一个新的衡量标准,以评估几个经过预先培训的大型模型(OPUS-MT、M2M-100)在翻译过程中对使用环境来纠正其偏见的敏感性。我们的方法不需要微调,因此在生产系统中可以很容易地使用,从陈规定型的性别职业偏见中去偏见翻译。