As the emerging trend of graph-based deep learning, Graph Neural Networks (GNNs) excel for their capability to generate high-quality node feature vectors (embeddings). However, the existing one-size-fits-all GNN implementations are insufficient to catch up with the evolving GNN architectures, the ever-increasing graph sizes, and the diverse node embedding dimensionalities. To this end, we propose \textbf{GNNAdvisor}, an adaptive and efficient runtime system to accelerate various GNN workloads on GPU platforms. First, GNNAdvisor explores and identifies several performance-relevant features from both the GNN model and the input graph, and uses them as a new driving force for GNN acceleration. Second, GNNAdvisor implements a novel and highly-efficient 2D workload management, tailored for GNN computation to improve GPU utilization and performance under different application settings. Third, GNNAdvisor capitalizes on the GPU memory hierarchy for acceleration by gracefully coordinating the execution of GNNs according to the characteristics of the GPU memory structure and GNN workloads. Furthermore, to enable automatic runtime optimization, GNNAdvisor incorporates a lightweight analytical model for an effective design parameter search. Extensive experiments show that GNNAdvisor outperforms the state-of-the-art GNN computing frameworks, such as Deep Graph Library ($3.02\times$ faster on average) and NeuGraph (up to $4.10\times$ faster), on mainstream GNN architectures across various datasets.
翻译:随着基于图形的深层学习的新兴趋势,图形神经网络(GNNS)在生成高质量节点特性矢量(组合)的能力方面表现优异。然而,现有的一刀切通用GNN实施系统不足以跟上GNN结构的演变、日益增长的图形尺寸和不同节点嵌入维度。为此,我们建议采用适应性高效运行时间系统,以加速GPU平台上的各种GNNN工作量。首先,GNNAdvisor探索和确定GNNN模式和输入图中的若干与业绩相关的特点,并把它们用作GNNNN加速的新驱动力。第二,GNNAvisor实施新的高效2D工作量管理,为GNN的计算而专门设计,以提高GPU在不同应用环境下的利用率和性能。第三,GNNAdvisor利用GPNM(通过优雅地协调GNNNS执行G的更快速度,在GPNMA的深度存储和GNNNG的模型模型中,使G的高级数字模型结构能够运行到G的GNNNS格式的模拟搜索。此外,使G的GSAS级的模型升级升级的模型升级的模型升级的模型升级升级升级的G-SDNNNTA升级升级的模型升级的模型升级的模型升级升级升级的模型升级的模型升级升级。