With the increasing demands for accountability, interpretability is becoming an essential capability for real-world AI applications. However, most methods utilize post-hoc approaches rather than training the interpretable model. In this article, we propose a novel interpretable model based on the concept bottleneck model (CBM). CBM uses concept labels to train an intermediate layer as the additional visible layer. However, because the number of concept labels restricts the dimension of this layer, it is difficult to obtain high accuracy with a small number of labels. To address this issue, we integrate supervised concepts with unsupervised ones trained with self-explaining neural networks (SENNs). By seamlessly training these two types of concepts while reducing the amount of computation, we can obtain both supervised and unsupervised concepts simultaneously, even for large-sized images. We refer to the proposed model as the concept bottleneck model with additional unsupervised concepts (CBM-AUC). We experimentally confirmed that the proposed model outperformed CBM and SENN. We also visualized the saliency map of each concept and confirmed that it was consistent with the semantic meanings.
翻译:由于对问责制的要求日益增加,可解释性正在成为现实世界的AI应用的一种基本能力。然而,大多数方法都采用后热处理方法,而不是培训可解释的模式。在本条中,我们提出一个基于概念瓶颈模型的新颖的可解释模式。CBM使用概念标签来训练中间层作为额外的可见层。然而,由于概念标签的数量限制了这一层的维度,因此很难以少量标签获得高度准确性。为了解决这个问题,我们把受监督的概念与受过自我解释神经网络(SENNs)培训的不受监督的概念结合起来。我们通过在减少计算数量的同时对这两种概念进行无缝的培训,我们可以同时获得既受监督的概念,又不受监督的概念,即使是大型图像。我们把拟议的模型称为带有其他不受监督的概念的瓶颈模型(CBM-AUSC)。我们实验性地确认,拟议的模型已经超越了CBM和SENN。我们还将每个概念的突出特征地图进行视觉化,并确认它与精准的含义是一致的。