Increasing utilization of machine learning based decision support systems emphasizes the need for resulting predictions to be both accurate and fair to all stakeholders. In this work we present a novel approach to increase a Neural Network model's fairness during training. We introduce a family of fairness enhancing regularization components that we use in conjunction with the traditional binary-cross-entropy based accuracy loss. These loss functions are based on Bias Parity Score (BPS), a score that helps quantify bias in the models with a single number. In the current work we investigate the behavior and effect of these regularization components on bias. We deploy them in the context of a recidivism prediction task as well as on a census-based adult income dataset. The results demonstrate that with a good choice of fairness loss function we can reduce the trained model's bias without deteriorating accuracy even in unbalanced dataset.
翻译:越来越多的利用基于机器学习的决策支持系统强调,由此得出的预测必须准确和公平,对所有利害相关者都是如此。在这项工作中,我们提出了一个新颖的方法,以提高神经网络模型在培训期间的公平性。我们引入了一个公平性家庭,加强正规化部分,我们与传统的双倍交叉机体准确性损失一起使用。这些损失功能基于“双倍均等分”(BPS),这一分数有助于用单一数字量化模型中的偏见。在目前的工作中,我们调查这些正规化部分的行为和对偏见的影响。我们将这些部分用于累犯预测任务和基于普查的成人收入数据集。结果表明,如果选择良好的公平性损失功能,我们可以减少经过培训的模式的偏差,即使数据组的准确性也不会下降。