Backpropagation-based visualizations have been proposed to interpret convolutional neural networks (CNNs), however a theory is missing to justify their behaviors: Guided backpropagation (GBP) and deconvolutional network (DeconvNet) generate more human-interpretable but less class-sensitive visualizations than saliency map. Motivated by this, we develop a theoretical explanation revealing that GBP and DeconvNet are essentially doing (partial) image recovery which is unrelated to the network decisions. Specifically, our analysis shows that the backward ReLU introduced by GBP and DeconvNet, and the local connections in CNNs are the two main causes of compelling visualizations. Extensive experiments are provided that support the theoretical analysis.
翻译:为了解释进化神经网络(CNNs),人们提议了基于后推进的视觉化,然而,却缺少一个理论来为其行为辩护:向后推进(GBP)和向后推进(DeconvNet)产生比突出的图象更人性化但对阶级敏感程度更低的视觉化。为此,我们开发了一个理论解释,揭示了英镑和DeconvNet基本上正在(部分)恢复与网络决定无关的图像。具体地说,我们的分析表明,由英镑和DeconvNet推出的后向ReLU以及CNN的本地连接是令人信服的视觉化的两大原因。我们提供了广泛的实验来支持理论分析。