With graphs rapidly growing in size and deeper graph neural networks (GNNs) emerging, the training and inference of GNNs become increasingly expensive. Existing network weight pruning algorithms cannot address the main space and computational bottleneck in GNNs, caused by the size and connectivity of the graph. To this end, this paper first presents a unified GNN sparsification (UGS) framework that simultaneously prunes the graph adjacency matrix and the model weights, for effectively accelerating GNN inference on large-scale graphs. Leveraging this new tool, we further generalize the recently popular lottery ticket hypothesis to GNNs for the first time, by defining a graph lottery ticket (GLT) as a pair of core sub-dataset and sparse sub-network, which can be jointly identified from the original GNN and the full dense graph by iteratively applying UGS. Like its counterpart in convolutional neural networks, GLT can be trained in isolation to match the performance of training with the full model and graph, and can be drawn from both randomly initialized and self-supervised pre-trained GNNs. Our proposal has been experimentally verified across various GNN architectures and diverse tasks, on both small-scale graph datasets (Cora, Citeseer and PubMed), and large-scale datasets from the challenging Open Graph Benchmark (OGB). Specifically, for node classification, our found GLTs achieve the same accuracies with 20%~98% MACs saving on small graphs and 25%~85% MACs saving on large ones. For link prediction, GLTs lead to 48%~97% and 70% MACs saving on small and large graph datasets, respectively, without compromising predictive performance. Codes available at https://github.com/VITA-Group/Unified-LTH-GNN.
翻译:随着图形在大小和深度图形神经网络(GNNS)中迅速增长,GNNS的培训和推断变得越来越昂贵。现有的网络重量调整算法无法解决GNNS中的主要空间和计算瓶颈问题,因为图形的大小和连通性导致GNNS中的主要空间和计算瓶颈问题。为此,本文件首先提出了一个统一的GNNS净化(UGS)框架,同时将图形的相近矩阵和模型重量牵引出来,以有效加快GNNN在大型图表中的误判。利用这一新工具,我们进一步将最近流行的彩票假设推广到GNNNNS,第一次将GLT彩票(GLT)定义为核心子数据集和零星子网络的组合。这个框架可以通过迭接地应用UGS。GNNN和全密度图形网络中的对应方程式一样,GLTT(GLT)可以进行远程培训,以便让GNNNG的精度数据和GNDOG的大规模基数据运行(OLDG),在O前的GLD GNLDG数据上随机初始和图中,在GGLDODODO的大规模数据上都在GODODODODODO中,在GLD中,在GLLD数据上都的大规模数据上,在GLDODODO值前的大规模数据上都没有找到。