Deep learning models have shown their superior performance in various vision tasks. However, the lack of precisely interpreting kernels in convolutional neural networks (CNNs) is becoming one main obstacle to wide applications of deep learning models in real scenarios. Although existing interpretation methods may find certain visual patterns which are associated with the activation of a specific kernel, those visual patterns may not be specific or comprehensive enough for interpretation of a specific activation of kernel of interest. In this paper, a simple yet effective optimization method is proposed to interpret the activation of any kernel of interest in CNN models. The basic idea is to simultaneously preserve the activation of the specific kernel and suppress the activation of all other kernels at the same layer. In this way, only visual information relevant to the activation of the specific kernel is remained in the input. Consistent visual information from multiple modified inputs would help users understand what kind of features are specifically associated with specific kernel. Comprehensive evaluation shows that the proposed method can help better interpret activation of specific kernels than widely used methods, even when two kernels have very similar activation regions from the same input image.
翻译:深层学习模型显示了其在各种视觉任务方面的优异表现,然而,在进化神经网络中缺乏精确的内核解释正在成为在现实情景中广泛应用深层学习模型的一个主要障碍。虽然现有的解释方法可能发现某些与激活特定内核有关的视觉模式,但这些视觉模式可能不够具体或全面,不足以解释具体激活感兴趣的内核。在本文中,提议一种简单而有效的优化方法来解释激活CNN模型中任何感兴趣的内核。基本想法是同时保存特定内核的激活,并抑制在同一层中所有其他内核的激活。这样,只有与激活特定内核有关的视觉信息才能留在输入中。从多处修改后输入的一致的视觉信息有助于用户了解具体内核的特性。全面评价表明,拟议的方法可以帮助更好地解释特定内核的激活,而不是广泛使用的方法,即使两个内核从同一输入图像中具有非常相似的激活区域。