Currently, there is a surge of interest in fair Artificial Intelligence (AI) and Machine Learning (ML) research which aims to mitigate discriminatory bias in AI algorithms, e.g. along lines of gender, age, and race. While most research in this domain focuses on developing fair AI algorithms, in this work, we show that a fair AI algorithm on its own may be insufficient to achieve its intended results in the real world. Using career recommendation as a case study, we build a fair AI career recommender by employing gender debiasing machine learning techniques. Our offline evaluation showed that the debiased recommender makes fairer career recommendations without sacrificing its accuracy. Nevertheless, an online user study of more than 200 college students revealed that participants on average prefer the original biased system over the debiased system. Specifically, we found that perceived gender disparity is a determining factor for the acceptance of a recommendation. In other words, our results demonstrate we cannot fully address the gender bias issue in AI recommendations without addressing the gender bias in humans.
翻译:目前,人们对公平人工智能(AI)和机器学习(ML)研究的兴趣激增,这些研究旨在减少AI算法中的歧视性偏见,例如性别、年龄和种族方面的歧视性偏见。虽然这一领域的大多数研究侧重于制定公平的AI算法,但在这项工作中,我们发现公平的AI算法本身可能不足以在现实世界实现预期的结果。利用职业建议作为案例研究,我们通过使用性别贬低机器学习技术来建立公平的AI职业建议。我们的离线评估显示,被贬低的推荐者在不牺牲准确性的情况下提出了更公平的职业建议。然而,对200多所大学生的在线用户研究表明,参与者平均倾向于最初的偏向系统,而不是被贬低的系统。具体地说,我们发现,认为性别差异是接受建议的决定性因素。换句话说,我们的结果表明,如果不解决人类的性别偏见问题,我们就无法在AI建议中充分解决性别偏见问题。