Explainable artificial intelligence (XAI) methods are currently evaluated with approaches mostly originated in interpretable machine learning (IML) research that focus on understanding models such as comparison against existing attribution approaches, sensitivity analyses, gold set of features, axioms, or through demonstration of images. There are problems with these methods such as that they do not indicate where current XAI approaches fail to guide investigations towards consistent progress of the field. They do not measure accuracy in support of accountable decisions, and it is practically impossible to determine whether one XAI method is better than the other or what the weaknesses of existing models are, leaving researchers without guidance on which research questions will advance the field. Other fields usually utilize ground-truth data and create benchmarks. Data representing ground-truth explanations is not typically used in XAI or IML. One reason is that explanations are subjective, in the sense that an explanation that satisfies one user may not satisfy another. To overcome these problems, we propose to represent explanations with canonical equations that can be used to evaluate the accuracy of XAI methods. The contributions of this paper include a methodology to create synthetic data representing ground-truth explanations, three data sets, an evaluation of LIME using these data sets, and a preliminary analysis of the challenges and potential benefits in using these data to evaluate existing XAI approaches. Evaluation methods based on human-centric studies are outside the scope of this paper.
翻译:目前,对可解释的人工智能(XAI)方法的评价主要来自可解释的机器学习(IML)研究,这些研究侧重于对模型的理解,例如与现有归属方法的比较、敏感性分析、黄金特征集、轴心或图像演示;这些方法存在问题,例如,这些方法没有表明当前XAI方法在哪些方面未能指导调查工作取得一致的实地进展;这些方法没有衡量支持问责决定的准确性,而且实际上不可能确定一种XAI方法是否优于其他方法,或现有模型的弱点,使研究人员对哪些研究问题将推进实地工作没有指导;其他领域通常使用地面真相数据并创建基准;代表地面真相解释的数据通常在XAI或IML中并不使用;其中一个原因是解释是主观的,因为一个用户无法满足另一个用户的要求;为了克服这些问题,我们提议用可用来评价XAI方法的准确性方程式进行解释;本文的贡献包括用一种方法来创建反映地面真相解释的合成数据,以及建立基准;三个数据解释方法用于外部分析。