AI explainability seeks to increase the transparency of models, making them more trustworthy in the process. The need for transparency has been recently motivated by the emergence of deep learning models, which are particularly obscure by nature. Even in the domain of images, where deep learning has succeeded the most, explainability is still poorly assessed. Multiple feature attribution methods have been proposed in the literature with the purpose of explaining a DL model's behavior using visual queues, but no standardized metrics to assess or select these methods exist. In this paper we propose a novel evaluation metric -- the Focus -- designed to quantify the faithfulness of explanations provided by feature attribution methods, such as LRP or GradCAM. First, we show the robustness of the metric through randomization experiments, and then use Focus to evaluate and compare three popular explainability techniques using multiple architectures and datasets. Our results find LRP and GradCAM to be consistent and reliable, the former being more accurate for high performing models, while the latter remains most competitive even when applied to poorly performing models. Finally, we identify a strong relation between Focus and factors like model architecture and task, unveiling a new unsupervised approach for the assessment of models.
翻译:AI 可解释性试图提高模型的透明度,使模型在这个过程中更加可信。透明度的需要最近是由于出现了深层次的学习模型,这些模型在性质上特别模糊。即使在深层次的学习最成功,对可解释性的评估仍然很差。文献中提出了多种特性归属方法,目的是用视觉队列解释DL模型的行为,但没有标准化的衡量标准来评估或选择这些方法。在本文件中,我们提出了一个新的评价标准 -- -- 焦点 -- -- 旨在量化特征归属方法(如LRP或GradCAM)所提供的解释的准确性。首先,我们通过随机化实验来显示该计量的稳健性,然后利用Focus来评估和比较三种通用解释技术。我们的结果发现LRP和GradCAM具有一致性和可靠性,而前者对于高性模型则更为准确,而后者即使在应用到表现不佳的模式时也仍然最具竞争力。最后,我们确定了焦点与模型结构和任务等要素之间的强有力关系,我们为模型的评估展示了一种新的不受监督的方法。