Today we have a good theoretical understanding of the representational power of Graph Neural Networks (GNNs). For example, their limitations have been characterized in relation to a hierarchy of Weisfeiler-Lehman (WL) isomorphism tests. However, we do not know what is encoded in the learned representations. This is our main question. We answer it using a probing framework to quantify the amount of meaningful information captured in graph representations. Our findings on molecular datasets show the potential of probing for understanding the inductive biases of graph-based models. We compare different families of models and show that transformer-based models capture more chemically relevant information compared to models based on message passing. We also study the effect of different design choices such as skip connections and virtual nodes. We advocate for probing as a useful diagnostic tool for evaluating graph-based models.
翻译:今天,我们对图形神经网络(GNNs)的代表性力量有了良好的理论理解。例如,它们的局限性在Weisfeiler-Lehman(WL)等式测试的等级方面已经具有了特征。然而,我们不知道在所学的表述中编码了什么。这是我们的主要问题。我们用一个验证框架来回答它,以量化图形表达中收集的有意义的信息的数量。我们对分子数据集的研究结果表明,在理解基于图形模型的进化偏差方面,有潜力。我们比较了不同的模型系列,并表明基于变压器的模型与基于信息传递的模型相比,收集了更化学上相关的信息。我们还研究了诸如跳过连接和虚拟节点等不同设计选择的影响。我们主张将验证作为评估基于图形模型的有用诊断工具。</s>