Uncovering rationales behind predictions of graph neural networks (GNNs) has received increasing attention over recent years. Instance-level GNN explanation aims to discover critical input elements, like nodes or edges, that the target GNN relies upon for making predictions. Though various algorithms are proposed, most of them formalize this task by searching the minimal subgraph which can preserve original predictions. However, an inductive bias is deep-rooted in this framework: several subgraphs can result in the same or similar outputs as the original graphs. Consequently, they have the danger of providing spurious explanations and fail to provide consistent explanations. Applying them to explain weakly-performed GNNs would further amplify these issues. To address this problem, we theoretically examine the predictions of GNNs from the causality perspective. Two typical reasons of spurious explanations are identified: confounding effect of latent variables like distribution shift, and causal factors distinct from the original input. Observing that both confounding effects and diverse causal rationales are encoded in internal representations, we propose a simple yet effective countermeasure by aligning embeddings. Concretely, concerning potential shifts in the high-dimensional space, we design a distribution-aware alignment algorithm based on anchors. This new objective is easy to compute and can be incorporated into existing techniques with no or little effort. Theoretical analysis shows that it is in effect optimizing a more faithful explanation objective in design, which further justifies the proposed approach.
翻译:最近几年来,人们越来越关注图形神经网络预测背后的隐蔽理由; 平面GNN的解释旨在发现关键输入要素,如节点或边缘,而GNN的预测则依赖于这些关键输入要素。 尽管提出了各种算法,但大多数算法都通过搜索能够保存原始预测的最小子图将这项任务正规化。然而,在这个框架中,隐含偏差根深蒂固:一些子图可以产生与原始图表相同或类似的产出。因此,它们有提供虚假解释的危险,无法提供一致的解释。运用这些解释来解释薄弱的GNNN的弱点,将进一步扩大这些问题。为解决这一问题,我们理论上从因果关系的角度审查了GNNN的预测。找出了两个典型的推论解释理由:潜在变量的影响,如分配变化,以及不同于原始投入的因果关系因素。观察,在内部陈述中,对各种因果关系的解释都有相互纠结的效果和不同的原因。 因此,我们建议一种简单有效的反制措施,通过将目标、不完善的GNNNNNN进一步加以解释来解释; 为了解决这个问题,我们从理论上从因果关系的角度来审视GNNNNNNNP的预测,在设计上进行更容易的调整。