Artificial intelligence algorithms are increasingly adopted as decisional aides by public organisations, with the promise of overcoming biases of human decision-makers. At the same time, the use of algorithms may introduce new biases in the human-algorithm interaction. A key concern emerging from psychology studies regards human overreliance on algorithmic advice even in the face of warning signals and contradictory information from other sources (automation bias). A second concern regards decision-makers inclination to selectively adopt algorithmic advice when it matches their pre-existing beliefs and stereotypes (selective adherence). To date, we lack rigorous empirical evidence about the prevalence of these biases in a public sector context. We assess these via two pre-registered experimental studies (N=1,509), simulating the use of algorithmic advice in decisions pertaining to the employment of school teachers in the Netherlands. In study 1, we test automation bias by exploring participants adherence to a prediction of teachers performance, which contradicts additional evidence, while comparing between two types of predictions: algorithmic v. human-expert. We do not find evidence for automation bias. In study 2, we replicate these findings, and we also test selective adherence by manipulating the teachers ethnic background. We find a propensity for adherence when the advice predicts low performance for a teacher of a negatively stereotyped ethnic minority, with no significant differences between algorithmic and human advice. Overall, our findings of selective, biased adherence belie the promise of neutrality that has propelled algorithm use in the public sector.
翻译:公共组织越来越多地采用人工智能算法作为决策助手,公共组织承诺克服人类决策者的偏见。与此同时,使用算法可能会在人类-人工互动中引入新的偏见。心理学研究中出现的一个关键关切是,即使在面临警告信号和其他来源的相互矛盾信息的情况下,人类还是过度依赖算法建议(烟雾偏差)。第二个关切是决策者倾向于选择性地采纳算法建议,以与其先前存在的信仰和陈规定型观念相匹配(选择性遵守)。迄今为止,我们缺乏关于公共部门中这些偏见普遍存在的严格经验证据。我们通过两项预先登记的实验研究(N=1,509)评估这些差异。我们在荷兰学校教师就业的决策中模拟使用算法建议。在研究1中,我们测试自动化偏重,探索参与者是否坚持对教师业绩的预测,这与其他证据相矛盾,同时比较两类预测:算法与人类专家(选择性遵守)。我们没有找到自动化偏差的证据。在研究2中,我们复制这些发现这些结果,我们还测试有选择性地遵守逻辑建议,在操纵教师的种族-种族-种族-偏差时,我们测试有选择性地预测教师的偏差。