The rising popularity of explainable artificial intelligence (XAI) to understand high-performing black boxes, also raised the question of how to evaluate explanations of machine learning (ML) models. While interpretability and explainability are often presented as a subjectively validated binary property, we consider it a multi-faceted concept. We identify 12 conceptual properties, such as Compactness and Correctness, that should be evaluated for comprehensively assessing the quality of an explanation. Our so-called Co-12 properties serve as categorization scheme for systematically reviewing the evaluation practice of more than 300 papers published in the last 7 years at major AI and ML conferences that introduce an XAI method. We find that 1 in 3 papers evaluate exclusively with anecdotal evidence, and 1 in 5 papers evaluate with users. We also contribute to the call for objective, quantifiable evaluation methods by presenting an extensive overview of quantitative XAI evaluation methods. This systematic collection of evaluation methods provides researchers and practitioners with concrete tools to thoroughly validate, benchmark and compare new and existing XAI methods. This also opens up opportunities to include quantitative metrics as optimization criteria during model training in order to optimize for accuracy and interpretability simultaneously.
翻译:可以解释的人工智能(XAI)越来越受欢迎,以理解高性能黑盒,这也提出了如何评价机器学习模型的解释的问题。虽然可解释性和可解释性常常作为主观验证的二元财产提出,但我们认为这是一个多面概念。我们确定了12个概念属性,如Clatical and correctity,为全面评估解释质量而应加以评估。我们所谓的Co-12属性是系统审查过去7年在采用XAI方法的主要AI和ML会议上发表的300多份文件的评价做法的分类办法。我们发现,每3份文件中就有1份文件只用传闻证据进行评价,每5份文件中就有1份文件与用户一道进行评价。我们还通过广泛概述XAI的定量评价方法,帮助呼吁采用客观、可量化的评价方法。这种系统收集的评价方法为研究人员和从业人员提供了具体工具,以便彻底验证、基准和比较新的和现有的XAI方法。这也为在示范培训期间将量化指标作为优化标准提供了机会,以便同时优化准确性和可解释性提供了。