In AI-assisted decision-making, a central promise of putting a human in the loop is that they should be able to complement the AI system by adhering to its correct and overriding its mistaken recommendations. In practice, however, we often see that humans tend to over- or under-rely on AI recommendations, meaning that they either adhere to wrong or override correct recommendations. Such reliance behavior is detrimental to decision-making accuracy. In this work, we articulate and analyze the interdependence between reliance behavior and accuracy in AI-assisted decision-making, which has been largely neglected in prior work. We also propose a visual framework to make this interdependence more tangible. This framework helps us interpret and compare empirical findings, as well as obtain a nuanced understanding of the effects of interventions (e.g., explanations) in AI-assisted decision-making. Finally, we infer several interesting properties from the framework: (i) when humans under-rely on AI recommendations, there may be no possibility for them to complement the AI in terms of decision-making accuracy; (ii) when humans cannot discern correct and wrong AI recommendations, no such improvement can be expected either; (iii) interventions may lead to an increase in decision-making accuracy that is solely driven by an increase in humans' adherence to AI recommendations, without any ability to discern correct and wrong. Our work emphasizes the importance of measuring and reporting both effects on accuracy and reliance behavior when empirically assessing interventions.
翻译:在AI辅助决策中,将人类放入环节的一个核心承诺是他们应该能够通过遵守其正确和改正其错误的建议来补充AI系统。然而,在实践中,我们经常看到人类倾向于过度或不足地依赖AI建议,这意味着他们要么遵循错误的建议,要么否决正确的建议。这种依赖行为对决策准确性有害。在本文中,我们阐述和分析了AI辅助决策中依赖行为和准确性之间的相互依存关系,这在先前的工作中被大部分忽略了。我们还提出了一种视觉框架,使这种相互依存关系更加具体化。这个框架有助于我们解释和比较实证研究结果,以及获得对AI辅助决策干预(例如解释)效果的细致理解。最后,我们从框架中推断出几个有趣的属性:(i)当人类过度依赖AI建议时,他们可能无法通过决策准确性方面来补充AI; (ii)当人类无法区分正确和错误的AI建议时,也不能期望有任何改善; (iii)干预措施可能导致决策准确性的增加,但这仅仅是由于人类遵循AI建议增加,而没有能力区分正确和错误。我们的工作强调了在实证评估干预措施时同时测量和报告对准确性和依赖行为的影响的重要性。