Uncovering rationales behind predictions of graph neural networks (GNNs) has received increasing attention over recent years. Instance-level GNN explanation aims to discover critical input elements, like nodes or edges, that the target GNN relies upon for making predictions. %These identified sub-structures can provide interpretations of GNN's behavior. Though various algorithms are proposed, most of them formalize this task by searching the minimal subgraph which can preserve original predictions. However, an inductive bias is deep-rooted in this framework: several subgraphs can result in the same or similar outputs as the original graphs. Consequently, they have the danger of providing spurious explanations and failing to provide consistent explanations. Applying them to explain weakly-performed GNNs would further amplify these issues. To address this problem, we theoretically examine the predictions of GNNs from the causality perspective. Two typical reasons for spurious explanations are identified: confounding effect of latent variables like distribution shift, and causal factors distinct from the original input. Observing that both confounding effects and diverse causal rationales are encoded in internal representations, \tianxiang{we propose a new explanation framework with an auxiliary alignment loss, which is theoretically proven to be optimizing a more faithful explanation objective intrinsically. Concretely for this alignment loss, a set of different perspectives are explored: anchor-based alignment, distributional alignment based on Gaussian mixture models, mutual-information-based alignment, etc. A comprehensive study is conducted both on the effectiveness of this new framework in terms of explanation faithfulness/consistency and on the advantages of these variants.
翻译:图形神经网络(GNNs)预测背后的未知理由近年来日益受到越来越多的关注。 实中GNN的解释旨在发现关键输入要素, 如节点或边缘, 目标GNN的预测依赖于这些关键输入要素。 % 这些查明的子结构可以解释GNN的行为。 虽然提出了各种算法, 多数算法通过搜索原始预测能够保存原始预测的最小子图将这项任务正规化。 但是, 感知偏差在这个框架中根深蒂固: 几个子集可以产生与原始图表相同或类似的输出。 因此, 它们有提供虚假解释和不提供一致解释的危险。 应用它们来解释低效GNNNs的次级结构可以进一步扩展这些问题。 为了解决这一问题,我们从理论上从因果关系的角度研究了GNN的预测。 找出了两个典型的解释理由: 潜在变量(如分配变化)的粘合效应, 和与原始投入不同的因果因素。 观察新的影响和不同的因果原理( ) 是全面解释性解释的复合性解释, 这些精确性解释在内部解释中, 以正统性解释为正统的逻辑解释。