Scalability of graph neural networks remains one of the major challenges in graph machine learning. Since the representation of a node is computed by recursively aggregating and transforming representation vectors of its neighboring nodes from previous layers, the receptive fields grow exponentially, which makes standard stochastic optimization techniques ineffective. Various approaches have been proposed to alleviate this issue, e.g., sampling-based methods and techniques based on pre-computation of graph filters. In this paper, we take a different approach and propose to use graph coarsening for scalable training of GNNs, which is generic, extremely simple and has sublinear memory and time costs during training. We present extensive theoretical analysis on the effect of using coarsening operations and provides useful guidance on the choice of coarsening methods. Interestingly, our theoretical analysis shows that coarsening can also be considered as a type of regularization and may improve the generalization. Finally, empirical results on real world datasets show that, simply applying off-the-shelf coarsening methods, we can reduce the number of nodes by up to a factor of ten without causing a noticeable downgrade in classification accuracy.
翻译:图形神经网络的可缩放性仍然是图形机器学习的主要挑战之一。由于节点的表示方式是通过对前层相邻节点的代表矢量进行递归和转换来计算,因此,可接收字段成倍增长,使得标准的随机优化技术无效。为缓解这一问题,提出了各种办法,例如,基于图形过滤器预数的基于取样的方法和技术。在本文中,我们采取不同的做法,并提议使用图表粗缩法对GNS进行可缩放的培训,这是通用的、极其简单的,在培训期间有亚线内存和时间成本。我们就使用粗略操作的效果进行了广泛的理论分析,并为粗略方法的选择提供了有益的指导。有趣的是,我们的理论分析表明,粗略也可以被视为一种正规化类型,并可能改进一般化。最后,关于真实世界数据集的经验结果表明,只要采用离沙粒的粗略方法,我们就可以将节点的数量降低到10个系数,而不会造成明显的下行精确的分类。