Group bias in natural language processing tasks manifests as disparities in system error rates across texts authorized by different demographic groups, typically disadvantaging minority groups. Dataset balancing has been shown to be effective at mitigating bias, however existing approaches do not directly account for correlations between author demographics and linguistic variables, limiting their effectiveness. To achieve Equal Opportunity fairness, such as equal job opportunity without regard to demographics, this paper introduces a simple, but highly effective, objective for countering bias using balanced training. We extend the method in the form of a gated model, which incorporates protected attributes as input, and show that it is effective at reducing bias in predictions through demographic input perturbation, outperforming all other bias mitigation techniques when combined with balanced training.
翻译:自然语言处理任务中的群体偏向表现为不同人口群体(通常不利于少数群体)授权的文本之间系统误差率的差异,这表现为不同人口群体(通常不利于少数群体)授权的文本之间系统误差率的差异。数据组合平衡在减少偏差方面已证明是有效的,但现有方法并未直接说明作者人口统计与语言变量之间的相互关系,限制了它们的效力。为了实现平等机会,例如不分人口统计的平等就业机会,本文件提出了一个简单但非常有效的目标,即利用平衡培训来消除偏见。我们扩展了一种门外模式,将受保护的属性作为投入,并表明这种方法能够有效地通过人口投入的渗透来减少预测中的偏差,在与均衡培训相结合的情况下,优于所有其他减少偏差的技术。