Graph neural networks (GNNs), as the de-facto model class for representation learning on graphs, are built upon the multi-layer perceptrons (MLP) architecture with additional message passing layers to allow features to flow across nodes. While conventional wisdom largely attributes the success of GNNs to their advanced expressivity for learning desired functions on nodes' ego-graphs, we conjecture that this is \emph{not} the main cause of GNNs' superiority in node prediction tasks. This paper pinpoints the major source of GNNs' performance gain to their intrinsic generalization capabilities, by introducing an intermediate model class dubbed as P(ropagational)MLP, which is identical to standard MLP in training, and then adopt GNN's architecture in testing. Intriguingly, we observe that PMLPs consistently perform on par with (or even exceed) their GNN counterparts across ten benchmarks and different experimental settings, despite the fact that PMLPs share the same (trained) weights with poorly-performed MLP. This critical finding opens a door to a brand new perspective for understanding the power of GNNs, and allow bridging GNNs and MLPs for dissecting their generalization behaviors. As an initial step to analyze PMLP, we show its essential difference with MLP at infinite-width limit lies in the NTK feature map in the post-training stage. Moreover, though MLP and PMLP cannot extrapolate non-linear functions for extreme OOD data, PMLP has more freedom to generalize near the training support.
翻译:图形神经网络(GNNS)是用于在图形上进行演示学习的脱facto模型类,它建在多层显示器(MLP)结构上,并增加传递信息层,以便通过节点传递特性。虽然常规智慧在很大程度上将GNNs的成功归功于在节点自我图谱上学习所需功能的高级表达式,但我们推测这是GNNS在节点预测任务中优势的主要原因。本文指出GNNs业绩的主要来源是其内在的概括化能力,通过引入一个中间模型类,称为P(ropagational)MLP,在培训中与标准 MLPP相似,然后在测试中采用GNNP的架构。我们注意到PMP在10个基准和不同实验环境中始终以(或甚至超过)GNNNP的对应方之间,尽管PMPP在总MP上拥有相同的(训练)比重,在MPLP的超常数能力上,在初始阶段,这个关键的模型在GNNFML上打开了一个基本的路径,在GML上显示其基本的路径,在GMDL的自我定位上可以理解。