The need for Explainable AI is increasing with the development of deep learning. The saliency maps derived from convolutional neural networks generally fail in localizing with accuracy the image features justifying the network prediction. This is because those maps are either low-resolution as for CAM [Zhou et al., 2016], or smooth as for perturbation-based methods [Zeiler and Fergus, 2014], or do correspond to a large number of widespread peaky spots as for gradient-based approaches [Sundararajan et al., 2017, Smilkov et al., 2017]. In contrast, our work proposes to combine the information from earlier network layers with the one from later layers to produce a high resolution Class Activation Map that is competitive with the previous art in term of insertion-deletion faithfulness metrics, while outperforming it in term of precision of class-specific features localization.
翻译:随着深层学习的发展,对可解释的AI的需求正在增加。来自进化神经网络的显要地图通常无法准确地对网络预测的图像特征进行本地化。这是因为这些地图要么是CAM[Zhou等人,2016年]的低分辨率,要么是扰动法[Zeiler和Fergus,2014年]的顺畅,或者的确与大量以梯度为基础的方法(Sundararajan等人,2017年,Smilkov等人,2017年 )的广泛高峰点相对应。相比之下,我们的工作提议将早期网络层的信息与后层的信息结合起来,以产生高分辨率分类激活图,该图与先前的插入-排异性指标方面的艺术相比具有竞争力,同时在班级特定特征本地化的精确性方面超过了它。