Machine Translation systems can produce different types of errors, some of which are characterized as critical or catastrophic due to the specific negative impact that they can have on users. In this paper we focus on one type of critical error: added toxicity. We evaluate and analyze added toxicity when translating a large evaluation dataset (HOLISTICBIAS, over 472k sentences, covering 13 demographic axes) from English into 164 languages. An automatic toxicity evaluation shows that added toxicity across languages varies from 0% to 5%. The output languages with the most added toxicity tend to be low-resource ones, and the demographic axes with the most added toxicity include sexual orientation, gender and sex, and ability. We also perform human evaluation on a subset of 8 translation directions, confirming the prevalence of true added toxicity. We use a measurement of the amount of source contribution to the translation, where a low source contribution implies hallucination, to interpret what causes toxicity. Making use of the input attributions allows us to explain toxicity, because the source contributions significantly correlate with toxicity for 84% of languages studied. Given our findings, our recommendations to reduce added toxicity are to curate training data to avoid mistranslations, mitigate hallucination and check unstable translations.
翻译:机器翻译系统会出现各种误译,其中一些被归类为关键或灾难性错误,因为它们可能对用户产生负面影响。本文侧重于一种关键性错误:增加毒性。我们对大型评估数据集(HOLISTICBIAS,超过472k句子,涵盖13个人口统计学轴)从英文翻译为164种语言时添加毒性进行评估和分析。自动毒性评估显示,跨语言的附加毒性从0%到5%不等。附加毒性最严重的输出语言往往是低资源语言,附加毒性最严重的人口统计学轴包括性取向、性别和能力。我们还对8个翻译方向的子集进行人工评估,确认了确实存在附加毒性。我们使用源贡献量的测量方法来解释毒性,其中较低的源贡献量意味着产生了幻觉。充分利用输入归因使我们能够解释毒性,因为84%的研究语言中,源贡献显著地与毒性相关。鉴于我们的发现,我们建议采取措施,减少增加毒性,包括筛选训练数据以避免错误翻译、减轻幻觉和检查不稳定的翻译。