Explanations are hypothesized to improve human understanding of machine learning models and achieve a variety of desirable outcomes, ranging from model debugging to enhancing human decision making. However, empirical studies have found mixed and even negative results. An open question, therefore, is under what conditions explanations can improve human understanding and in what way. Using adapted causal diagrams, we provide a formal characterization of the interplay between machine explanations and human understanding, and show how human intuitions play a central role in enabling human understanding. Specifically, we identify three core concepts of interest that cover all existing quantitative measures of understanding in the context of human-AI decision making: task decision boundary, model decision boundary, and model error. Our key result is that without assumptions about task-specific intuitions, explanations may potentially improve human understanding of model decision boundary, but they cannot improve human understanding of task decision boundary or model error. To achieve complementary human-AI performance, we articulate possible ways on how explanations need to work with human intuitions. For instance, human intuitions about the relevance of features (e.g., education is more important than age in predicting a person's income) can be critical in detecting model error. We validate the importance of human intuitions in shaping the outcome of machine explanations with empirical human-subject studies. Overall, our work provides a general framework along with actionable implications for future algorithmic development and empirical experiments of machine explanations.
翻译:假设解释是为了提高人类对机器学习模型的理解,并实现各种理想结果,从模型调试到加强人类决策,从模型调试到增强人类决策等,都得出了好坏参半的结果。因此,一个未决问题是,在什么条件下解释可以提高人类的理解,以何种方式提高人类的理解。我们使用经调整的因果图表,对机器解释与人类理解之间的相互作用进行正式描述,并表明人类直觉如何在帮助人类理解方面发挥核心作用。具体地说,我们确定了三个令人感兴趣的核心概念,这些概念涵盖了在人类-AI决策中现有的理解量化计量标准:任务决定界限、模型决定界限和模型错误。我们的关键结果是,如果没有对特定任务直觉的假设,解释可能提高人类对模式决定界限的理解,但解释可能提高人类对模式界限或模型错误的理解。为了实现人类-AI的互补性表现,我们阐述了解释人类直觉需要如何与人类直觉相结合的可能方法。例如,人类对特征的相关性的直觉比年龄在预测一个人的收入方面更为重要。我们的主要结果是,我们的主要结果是,我们的主要结果是没有对模型分析,我们一般的逻辑分析,我们的工作研究会提供人类结果分析。