Women are often perceived as junior to their male counterparts, even within the same job titles. While there has been significant progress in the evaluation of gender bias in natural language processing (NLP), existing studies seldom investigate how biases toward gender groups change when compounded with other societal biases. In this work, we investigate how seniority impacts the degree of gender bias exhibited in pretrained neural generation models by introducing a novel framework for probing compound bias. We contribute a benchmark robustness-testing dataset spanning two domains, U.S. senatorship and professorship, created using a distant-supervision method. Our dataset includes human-written text with underlying ground truth and paired counterfactuals. We then examine GPT-2 perplexity and the frequency of gendered language in generated text. Our results show that GPT-2 amplifies bias by considering women as junior and men as senior more often than the ground truth in both domains. These results suggest that NLP applications built using GPT-2 may harm women in professional capacities.
翻译:虽然在评估自然语言处理过程中的性别偏见方面取得了重大进展,但现有的研究很少调查在与其他社会偏见相结合的情况下,对性别群体的偏见会如何改变;在这项工作中,我们调查资历如何影响在经过培训的神经生成模型中表现出的性别偏见的程度,方法是采用新的框架来检验复合偏见;我们协助建立一个基准强力测试数据集,该数据集涵盖两个领域:美国参议员和教授,使用远视方法创建。我们的数据集包括人文文字,其中含有基本事实和对立反事实。我们随后检查GPT-2的模糊性以及生成文本中性别语言的频率。我们的结果显示,GPT-2将妇女视为年幼和男子,在这两个领域都比实际真相更经常地视为年长,从而扩大了偏见。这些结果表明,利用GPT-2建立的NLP应用程序可能损害妇女的专业能力。