Recently, the Weisfeiler-Lehman (WL) graph isomorphism test was used to measure the expressiveness of graph neural networks (GNNs), showing that the neighborhood aggregation GNNs were at most as powerful as 1-WL test in distinguishing graph structures. There were also improvements proposed in analogy to $k$-WL test ($k>1$). However, the aggregators in these GNNs are far from injective as required by the WL test, and suffer from weak distinguishing strength, making it become expressive bottlenecks. In this paper, we improve the expressiveness by exploring powerful aggregators. We reformulate aggregation with the corresponding aggregation coefficient matrix, and then systematically analyze the requirements of the aggregation coefficient matrix for building more powerful aggregators and even injective aggregators. It can also be viewed as the strategy for preserving the rank of hidden features, and implies that basic aggregators correspond to a special case of low-rank transformations. We also show the necessity of applying nonlinear units ahead of aggregation, which is different from most aggregation-based GNNs. Based on our theoretical analysis, we develop two GNN layers, ExpandingConv and CombConv. Experimental results show that our models significantly boost performance, especially for large and densely connected graphs.
翻译:最近,Weisfeiler-Lehman(WL)图的形态测试被用来测量图形神经网络(GNNS)的清晰度,显示相邻聚合GNNS在不同的图形结构中最强大,其强度与1-WL测试最强。在类推美元-WL测试(k>1美元)中也提出了改进建议。然而,这些GNNS中的聚合器远未像WL测试所要求的那样进行注射,并且受到微弱的分辨力的影响,使它成为显性瓶颈。在本文中,我们通过探索强大的聚合器来改进图像的清晰度。我们重新配置相应的聚合系数矩阵,然后系统地分析聚合系数矩阵对建立更强大的聚合器甚至注射聚合器的要求。它也可以被视为保存隐藏特性等级的战略,意味着基本聚合器与低级变异的特殊案例相对应,因此它成为了表征的瓶颈。我们还表明,在聚合组群之前应用非线性单位的必要性,这与大多数基于集群的聚合系数矩阵矩阵矩阵矩阵矩阵,特别是GNNNNP和G的深度变现模型展示了我们G的深度变现。