Recently, Graph Neural Networks (GNNs) have received a lot of interest because of their success in learning representations from graph structured data. However, GNNs exhibit different compute and memory characteristics compared to traditional Deep Neural Networks (DNNs). Graph convolutions require feature aggregations from neighboring nodes (known as the aggregation phase), which leads to highly irregular data accesses. GNNs also have a very regular compute phase that can be broken down to matrix multiplications (known as the combination phase). All recently proposed GNN accelerators utilize different dataflows and microarchitecture optimizations for these two phases. Different communication strategies between the two phases have been also used. However, as more custom GNN accelerators are proposed, the harder it is to qualitatively classify them and quantitatively contrast them. In this work, we present a taxonomy to describe several diverse dataflows for running GNN inference on accelerators. This provides a structured way to describe and compare the design-space of GNN accelerators.
翻译:最近,图形神经网络(GNN)由于成功地从图形结构化数据中学习了表征,因此获得了很大的兴趣。然而,与传统的深神经网络(DNN)相比,GNN的计算和记忆特征不同。图表演变需要邻居节点(称为聚合阶段)的特征聚合(称为聚合阶段),从而导致极不规律的数据存取。GNN也有一个非常正常的计算阶段,可以细分为矩阵乘数(称为组合阶段)。所有最近提出的GNN加速器都利用了这两个阶段的不同数据流和微结构优化。这两个阶段还使用了不同的通信战略。然而,随着越来越多的定制的GNNN加速器被提议,对其进行定性分类和定量对比的难度越大。在这项工作中,我们提出了一个分类,用来描述在加速器上运行GNNN的多倍推法的多种数据流。这为描述和比较GNNN加速器的设计空间提供了一种结构化的方法。