While Graph Neural Networks (GNNs) have achieved remarkable results in a variety of applications, recent studies exposed important shortcomings in their ability to capture the structure of the underlying graph. It has been shown that the expressive power of standard GNNs is bounded by the Weisfeiler-Leman (WL) graph isomorphism test, from which they inherit proven limitations such as the inability to detect and count graph substructures. On the other hand, there is significant empirical evidence, e.g. in network science and bioinformatics, that substructures are often intimately related to downstream tasks. To this end, we propose "Graph Substructure Networks" (GSN), a topologically-aware message passing scheme based on substructure encoding. We theoretically analyse the expressive power of our architecture, showing that it is strictly more expressive than the WL test, and provide sufficient conditions for universality. Importantly, we do not attempt to adhere to the WL hierarchy; this allows us to retain multiple attractive properties of standard GNNs such as locality and linear network complexity, while being able to disambiguate even hard instances of graph isomorphism. We perform an extensive experimental evaluation on graph classification and regression tasks and obtain state-of-the-art results in diverse real-world settings including molecular graphs and social networks. The code is publicly available at https://github.com/gbouritsas/graph-substructure-networks.
翻译:虽然图形神经网络(GNNs)在各种应用中取得了显著成果,但最近的研究揭示了它们捕捉基本图形结构的能力存在重大缺陷。已经表明,标准的GNNs的表情力量受Weisfeiler-Leman(WL)图像的形态测试的约束,它们从中继承了已证明的局限性,例如无法探测和计数子结构。另一方面,在网络科学和生物信息学方面,有大量的经验证据表明,次级结构往往与下游任务密切相关。为此,我们提议“Graph 亚结构网络”(GSN),这是一个基于子结构编码的上层认知信息传递计划。我们从理论上分析我们架构的表情力量,表明它比WL测试更能表达,并为普遍性提供了充分的条件。重要的是,我们并不试图遵守WL等级;这使我们得以保留标准GNNS的多种有吸引力的特性,例如地点和线性网络复杂性。我们既能让数据库/线性网络变得不易懂,又能让数据库的内层结构结构化,同时在结构编码上进行广泛的实验性翻版和图状图状的模型的模型分析。我们进行广泛的分析。我们在正态上进行大量的图表和图式分析,在正态结构上进行不易入。