In recent years, graph neural networks (GNNs) have achieved great success in the field of graph representation learning. Although prior work has shed light into the expressiveness of those models (\ie whether they can distinguish pairs of non-isomorphic graphs), it is still not clear what structural information is encoded into the node representations that are learned by those models. In this paper, we investigate which properties of graphs are captured purely by these models, when no node attributes are available. Specifically, we study four popular GNN models, and we show that two of them embed all nodes into the same feature vector, while the other two models generate representations that are related to the number of walks over the input graph. Strikingly, structurally dissimilar nodes can have similar representations at some layer $k>1$, if they have the same number of walks of length $k$. We empirically verify our theoretical findings on real datasets.
翻译:近年来,图神经网络(GNNs)在图表示学习方面取得了巨大的成功。尽管先前的工作已经揭示了这些模型的表达能力(\ie是否能区分非同构图对),但仍不清楚学习的节点表示中编码了哪些结构信息,尤其在没有节点属性的情况下。本文研究了四种流行的GNN模型,并展示了其中两种将所有节点嵌入到相同的特征向量中,另外两种模型生成的表示与输入图上的行走次数有关。令人惊异的是,如果两个结构不同的节点在某个层$k>1$中有相同长度的行走数,它们可能具有相似的表示。我们在真实数据集上通过实验证实了我们的理论发现。