Despite AI's superhuman performance in a variety of domains, humans are often unwilling to adopt AI systems. The lack of interpretability inherent in many modern AI techniques is believed to be hurting their adoption, as users may not trust systems whose decision processes they do not understand. We investigate this proposition with a novel experiment in which we use an interactive prediction task to analyze the impact of interpretability and outcome feedback on trust in AI and on human performance in AI-assisted prediction tasks. We find that interpretability led to no robust improvements in trust, while outcome feedback had a significantly greater and more reliable effect. However, both factors had modest effects on participants' task performance. Our findings suggest that (1) factors receiving significant attention, such as interpretability, may be less effective at increasing trust than factors like outcome feedback, and (2) augmenting human performance via AI systems may not be a simple matter of increasing trust in AI, as increased trust is not always associated with equally sizable improvements in performance. These findings invite the research community to focus not only on methods for generating interpretations but also on techniques for ensuring that interpretations impact trust and performance in practice.
翻译:尽管大赦国际在许多领域表现超人,但人类往往不愿意采用大赦国际系统。许多现代大赦国际技术所固有的缺乏可解释性被认为不利于采用这些系统,因为用户可能不相信其决策程序不理解的系统。我们用一个新颖的实验来调查这一提议,我们用一个互动的预测任务来分析可解释性和结果反馈对大赦国际信任和大赦国际协助的预测任务中人类业绩的影响。我们发现,可解释性没有带来强有力的信任改善,而结果反馈的效果则大得多、更可靠。但是,这两个因素对参与者的任务业绩影响不大。我们的调查结果表明,(1) 受到极大关注的因素,例如可解释性,在增加信任方面可能不如结果反馈等因素那么有效,(2) 通过大赦国际系统提高人类业绩,可能不是简单的事情,因为增加信任并不总是与同样显著的业绩改进联系在一起。这些调查结果请研究界不仅注重产生解释的方法,而且注重确保解释影响信任和实践中业绩的技术。