Graph Neural Networks (GNNs) are proposed without considering the agnostic distribution shifts between training and testing graphs, inducing the degeneration of the generalization ability of GNNs on Out-Of-Distribution (OOD) settings. The fundamental reason for such degeneration is that most GNNs are developed based on the I.I.D hypothesis. In such a setting, GNNs tend to exploit subtle statistical correlations existing in the training set for predictions, even though it is a spurious correlation. However, such spurious correlations may change in testing environments, leading to the failure of GNNs. Therefore, eliminating the impact of spurious correlations is crucial for stable GNNs. To this end, we propose a general causal representation framework, called StableGNN. The main idea is to extract high-level representations from graph data first and resort to the distinguishing ability of causal inference to help the model get rid of spurious correlations. Particularly, we exploit a graph pooling layer to extract subgraph-based representations as high-level representations. Furthermore, we propose a causal variable distinguishing regularizer to correct the biased training distribution. Hence, GNNs would concentrate more on the stable correlations. Extensive experiments on both synthetic and real-world OOD graph datasets well verify the effectiveness, flexibility and interpretability of the proposed framework.
翻译:建议的神经网络(GNNs)没有考虑培训和测试图表之间的不可知分布变化,从而导致GNNs关于外向分配(OOOD)设置的普遍化能力退化。这种退化的根本原因是,大多数GNNs是根据I.I.D假设开发的。在这种背景下,GNNs往往利用一套预测培训中存在的微妙的统计相关性,尽管这是一种虚假的关联。然而,这种虚假的相关性可能会在测试环境中发生变化,导致GNNs的失败。因此,消除假相关系的影响对于GNNs的稳定GNs来说至关重要。为此,我们提议了一个总的因果代表框架,称为StablGNN。 其主要想法是从图形数据中提取高层次的表示,并首先利用因果推断能力来帮助模型消除虚假的关联。特别是,我们利用一个图形集合层来提取基于子图的演示,作为高层次的演示。此外,我们提议一个因果变量来区分虚假的对应关系,以校正的定期性分析常规性来纠正O型数据的真实性。因此,对GNNS的模型进行更精确的模型的模拟分析。