The interpretability of deep neural networks has attracted increasing attention in recent years, and several methods have been created to interpret the "black box" model. Fundamental limitations remain, however, that impede the pace of understanding the networks, especially the extraction of understandable semantic space. In this work, we introduce the framework of semantic explainable AI (S-XAI), which utilizes row-centered principal component analysis to obtain the common traits from the best combination of superpixels discovered by a genetic algorithm, and extracts understandable semantic spaces on the basis of discovered semantically sensitive neurons and visualization techniques. Statistical interpretation of the semantic space is also provided, and the concept of semantic probability is proposed for the first time. Our experimental results demonstrate that S-XAI is effective in providing a semantic interpretation for the CNN, and offers broad usage, including trustworthiness assessment and semantic sample searching.
翻译:近些年来,深神经网络的可解释性日益引起人们的注意,并制定了几种方法来解释“黑盒”模式,但仍存在一些根本限制,妨碍理解网络的速度,特别是难以理解的语义空间的提取。在这项工作中,我们引入了语义可解释的AI(S-XAI)框架,它利用以行为中心的主要组成部分分析,从遗传算法发现的超级像素的最佳组合中获取共同特征,并根据所发现的语义敏感神经元和可视化技术提取可理解的语义空间。还提供了语义空间的统计解释,首次提出了语义概率概念。我们的实验结果表明,S-XAI在为CNN提供语义解释方面是有效的,并且提供了广泛的用途,包括可信度评估和语义样本搜索。