Graph Neural Networks (GNNs) are based on repeated aggregations of information across nodes' neighbors in a graph. However, because common neighbors are shared between different nodes, this leads to repeated and inefficient computations. We propose Hierarchically Aggregated computation Graphs (HAGs), a new GNN graph representation that explicitly avoids redundancy by managing intermediate aggregation results hierarchically, eliminating repeated computations and unnecessary data transfers in GNN training and inference. We introduce an accurate cost function to quantitatively evaluate the runtime performance of different HAGs and use a novel HAG search algorithm to find optimized HAGs. Experiments show that the HAG representation significantly outperforms the standard GNN graph representation by increasing the end-to-end training throughput by up to 2.8x and reducing the aggregations and data transfers in GNN training by up to 6.3x and 5.6x, while maintaining the original model accuracy.
翻译:神经网络图(GNN)基于一个图中各节点邻居之间重复汇总的信息。 但是,由于共同邻居在不同节点之间共享信息,这会导致重复和低效率的计算。 我们提议了等级综合计算图(HAGs),这是一个新的GNN图形,通过按等级管理中间汇总结果,明确避免冗余的GNN图形,消除了在GNN培训和推论中重复计算和不必要的数据传输。我们引入了准确的成本功能,对不同 HAGs运行时间的性能进行定量评估,并使用新型的HAG搜索算法寻找优化的HAGs。 实验显示,HAG代表明显超过标准的GNN图形代表,将端到端培训的吞吐量增加至2.8x,并将GNN培训的汇总和数据传输减少至6.3x和5.6x,同时保持原始模型准确性。