Graph neural networks (GNNs) have become compelling models designed to perform learning and inference on graph-structured data, but little work has been done on understanding the fundamental limitations of GNNs to be scalable to larger graphs and generalized to out-of-distribution inputs. In this paper, we use a random graph generator that allows us to systematically investigate how the graph size and structural properties affect the predictive performance of GNNs. We present specific evidence that, among the many graph properties, the mean and modality of the node degree distribution are the key features that determine whether GNNs can generalize to unseen graphs. Accordingly, we propose flexible GNNs (Flex-GNNs), using multiple node update functions and the inner loop optimization as a generalization to the single type of canonical nonlinear transformation over aggregated inputs, allowing the network to adapt flexibly to new graphs. The Flex-GNN framework improves the generalization out of the training set on several inference tasks.
翻译:图形神经网络(GNNs)已成为令人信服的模型,旨在对图形结构数据进行学习和推断,但是在理解GNNs的基本局限性方面却没有做多少工作,这些基本局限性可以伸缩到更大的图表上,并被广泛推广到分布性输入中。在本文中,我们使用随机图形生成器,以便系统调查图形大小和结构属性如何影响GNS的预测性能。我们提出具体证据,说明在许多图形属性中,节点分布的平均值和模式是决定GNS能否向看不见的图表概括的关键特征。因此,我们提议采用多节点更新功能和内部循环优化,作为单一类型非线性非线性转换的概括,而不是总投入,使网络能够灵活地适应新的图表。Flex-GNNN框架改进了关于若干推论任务的培训的概括性。