The opaque reasoning of Graph Neural Networks induces a lack of human trust. Existing graph network explainers attempt to address this issue by providing post-hoc explanations, however, they fail to make the model itself more interpretable. To fill this gap, we introduce the Concept Encoder Module, the first differentiable concept-discovery approach for graph networks. The proposed approach makes graph networks explainable by design by first discovering graph concepts and then using these to solve the task. Our results demonstrate that this approach allows graph networks to: (i) attain model accuracy comparable with their equivalent vanilla versions, (ii) discover meaningful concepts that achieve high concept completeness and purity scores, (iii) provide high-quality concept-based logic explanations for their prediction, and (iv) support effective interventions at test time: these can increase human trust as well as significantly improve model performance.
翻译:图形神经网络的不透明推理导致人类缺乏信任。 现有的图形网络解释者试图通过提供热后解释来解决这一问题,但是,它们未能使模型本身更容易解释。 为了填补这一空白,我们引入了概念编码模块,这是图形网络第一个不同的概念发现方法。 拟议的方法使图形网络可以通过设计来解释,先发现图形概念,然后用这些概念来解决问题。 我们的结果表明,这一方法允许图形网络:(一) 获得与其等同的香草版本相类似的模型准确性,(二) 发现实现高概念完整性和纯度评分的有意义的概念,(三) 为预测提供高质量的基于概念的逻辑解释,(四) 支持测试时的有效干预:这些可以增加人类的信任,并大大改进模型性能。