The main objective of eXplainable Artificial Intelligence (XAI) is to provide effective explanations for black-box classifiers. The existing literature lists many desirable properties for explanations to be useful, but there is no consensus on how to quantitatively evaluate explanations in practice. Moreover, explanations are typically used only to inspect black-box models, and the proactive use of explanations as a decision support is generally overlooked. Among the many approaches to XAI, a widely adopted paradigm is Local Linear Explanations - with LIME and SHAP emerging as state-of-the-art methods. We show that these methods are plagued by many defects including unstable explanations, divergence of actual implementations from the promised theoretical properties, and explanations for the wrong label. This highlights the need to have standard and unbiased evaluation procedures for Local Linear Explanations in the XAI field. In this paper we address the problem of identifying a clear and unambiguous set of metrics for the evaluation of Local Linear Explanations. This set includes both existing and novel metrics defined specifically for this class of explanations. All metrics have been included in an open Python framework, named LEAF. The purpose of LEAF is to provide a reference for end users to evaluate explanations in a standardised and unbiased way, and to guide researchers towards developing improved explainable techniques.
翻译:现有文献列出了许多值得解释的属性,供解释有用,但对于如何从数量上评估解释在实践中没有共识。此外,解释通常仅用于检查黑箱模型,而主动使用解释作为决策支持通常被忽视。在XAI的许多做法中,广泛采用的范例是局部线性解释,其中LIME和SHAP是作为最新解释方法而出现的。我们表明,这些方法有许多缺陷,包括解释不稳定、实际执行与承诺的理论属性的差异以及错误标签的解释。这突出表明,需要为XAI字段的本地线性解释制定标准和公正的评价程序。在本文件中,我们处理为评估本地线性解释确定一套明确和毫不含糊的衡量标准的问题。这套标准包括专门为这一类解释界定的现有和新的衡量标准。所有指标都包含在开放的Python框架中,称为LEAFAF, 用于为研究人员制定可改进的最终解释方法提供一种标准性解释。LAAAF的目的是为最终评估提供一种可改进的方法。