Convolutional neural network (CNN) models have seen advanced improvements in performance in various domains, but lack of interpretability is a major barrier to assurance and regulation during operation for acceptance and deployment of AI-assisted applications. There have been many works on input interpretability focusing on analyzing the input-output relations, but the internal logic of models has not been clarified in the current mainstream interpretability methods. In this study, we propose a novel hybrid CNN-interpreter through: (1) An original forward propagation mechanism to examine the layer-specific prediction results for local interpretability. (2) A new global interpretability that indicates the feature correlation and filter importance effects. By combining the local and global interpretabilities, hybrid CNN-interpreter enables us to have a solid understanding and monitoring of model context during the whole learning process with detailed and consistent representations. Finally, the proposed interpretabilities have been demonstrated to adapt to various CNN-based model structures.
翻译:在各个领域,进化神经网络(CNN)模型的性能都有了进步,但缺乏可解释性是接受和部署AI辅助应用程序操作期间的保证和监管的主要障碍,许多关于投入可解释性的工作侧重于分析投入-产出关系,但目前的主流可解释性方法没有澄清模型的内部逻辑。在本研究报告中,我们建议采用一个新的CNN混合解释法,其方法是:(1) 一种原始的前方传播机制,以审查具体层次的预测结果,以了解当地可解释性。(2) 一种新的全球可解释性,表明特征的相关性和过滤重要性效应。通过将本地和全球的解释性相结合,CNN混合解释能使我们在整个学习过程中能够以详细和一致的表述对模型背景有扎实的了解和监测。最后,已经证明拟议的解释性能够适应CNN的各种模式结构。