Graph neural networks (GNNs) have shown great prowess in learning representations suitable for numerous graph-based machine learning tasks. When applied to semi-supervised node classification, GNNs are widely believed to work well due to the homophily assumption ("like attracts like"), and fail to generalize to heterophilous graphs where dissimilar nodes connect. Recent works design new architectures to overcome such heterophily-related limitations, citing poor baseline performance and new architecture improvements on a few heterophilous graph benchmark datasets as evidence for this notion. In our experiments, we empirically find that standard graph convolutional networks (GCNs) can actually achieve better performance than such carefully designed methods on some commonly used heterophilous graphs. This motivates us to reconsider whether homophily is truly necessary for good GNN performance. We find that this claim is not quite true, and in fact, GCNs can achieve strong performance on heterophilous graphs under certain conditions. Our work carefully characterizes these conditions, and provides supporting theoretical understanding and empirical observations. Finally, we examine existing heterophilous graphs benchmarks and reconcile how the GCN (under)performs on them based on this understanding.
翻译:在适合许多基于图形的机器学习任务的学习表现方面,GNNs 表现出了巨大的勇气。当应用于半监督节点分类时,由于同质假设(“像吸引像”一样),GNS被广泛认为效果良好,并且没有概括到不同节点连接的异同哲学图表。最近的工作设计了新结构,以克服这些与杂交相关的限制,以基准性能差和新结构的改进为根据,在少数偏激的图表基准数据集上作为这个概念的证据。我们实验中,我们从经验上发现标准图形共变网络(GCNs)实际上能够比一些常用的异异同性图形上精心设计的方法取得更好的效果。这促使我们重新考虑共性是否真正需要GNN的好表现。我们发现,这一说法并不十分正确,事实上,GCNs在某些条件下可以在异性图形上取得很强的性能。我们的工作仔细描述了这些条件,并且提供了支持理论理解和实验性观测结果。最后,我们根据现有的GPROBresmas的图表和理解,我们根据现有的GPIPImagrammas 来审视了这些基准。