Increasing demands for understanding the internal behaviors of convolutional neural networks (CNNs) have led to remarkable improvements in explanation methods. Particularly, several class activation mapping (CAM) based methods, which generate visual explanation maps by a linear combination of activation maps from CNNs, have been proposed. However, the majority of the methods lack a theoretical basis in how to assign their weighted linear coefficients. In this paper, we revisit the intrinsic linearity of CAM w.r.t. the activation maps. Focusing on the linearity, we construct an explanation model as a linear function of binary variables which denote the existence of the corresponding activation maps. With this approach, the explanation model can be determined by the class of additive feature attribution methods which adopts SHAP values as a unified measure of feature importance. We then demonstrate the efficacy of the SHAP values as the weight coefficients for CAM. However, the exact SHAP values are incalculable. Hence, we introduce an efficient approximation method, referred to as LIFT-CAM. On the basis of DeepLIFT, our proposed method can estimate the true SHAP values quickly and accurately. Furthermore, it achieves better performances than the other previous CAM-based methods in qualitative and quantitative aspects.
翻译:对理解进化神经网络(CNNs)内部行为的需求不断增加,导致解释方法的显著改进。特别是,提出了几种基于阶级激活映射(CAM)的方法,这些方法通过CNN的激活地图的线性组合产生直线解释地图;然而,大多数方法缺乏如何分配加权线性系数的理论基础。在本文件中,我们重新审视CAM w.r.t. 启动图的内在线性。我们以线性为重点,构建一个解释模型,作为显示相应激活地图存在的二进制变量的线性函数。采用这一方法,解释模型可以由将SHAP值作为特征重要性统一衡量标准的添加性特征属性归属方法类别来决定。然后,我们将SHAP值作为CAM加权系数的功效展示出来。然而,精确的SHAP值是无法估量的。因此,我们采用了一种高效的近似方法,称为LIFT-CAM。在DeepLIFT的基础上,我们提出的方法可以快速准确地估计真实的SHAP值。此外,我们提出的方法在质量方面比以前更好地评估了其他以SHAP为基础的定量方法。