Graph neural networks (GNNs) are highly effective on a variety of graph-related tasks; however, they lack interpretability and transparency. Current explainability approaches are typically local and treat GNNs as black-boxes. They do not look inside the model, inhibiting human trust in the model and explanations. Motivated by the ability of neurons to detect high-level semantic concepts in vision models, we perform a novel analysis on the behaviour of individual GNN neurons to answer questions about GNN interpretability, and propose new metrics for evaluating the interpretability of GNN neurons. We propose a novel approach for producing global explanations for GNNs using neuron-level concepts to enable practitioners to have a high-level view of the model. Specifically, (i) to the best of our knowledge, this is the first work which shows that GNN neurons act as concept detectors and have strong alignment with concepts formulated as logical compositions of node degree and neighbourhood properties; (ii) we quantitatively assess the importance of detected concepts, and identify a trade-off between training duration and neuron-level interpretability; (iii) we demonstrate that our global explainability approach has advantages over the current state-of-the-art -- we can disentangle the explanation into individual interpretable concepts backed by logical descriptions, which reduces potential for bias and improves user-friendliness.
翻译:在各种与图表有关的任务中,神经神经网络(GNNs)非常有效;但是,它们缺乏解释性和透明度。目前的解释性方法通常是局部的,并且将GNNs作为黑箱对待。它们不看模型内,抑制了人类对模型和解释的信任。受神经元在视觉模型中检测高层次语义概念的能力的驱动,我们对单个GNN神经元的行为进行了新颖的分析,以解答关于GNN可解释性的问题,提出了评估GNN神经元可解释性的新指标。我们提出了一个新颖的方法,利用神经级概念为GNNs提供全球解释性解释,使实践者能够对模型有一个高层次的视角。具体地说,(一)根据我们的知识,这是第一个工作,表明GNNN神经作为概念的检测器,与作为无度和邻里特性的逻辑构成而形成的概念非常一致;(二)我们量化地评估了所检测到的概念的重要性,并确定了在培训期间和神经级可解释性之间进行权衡性权衡的重要性。我们提出了一种新的方法,用以解释性地解释性解释性地解释我们所支持的逻辑性解释的每个概念。