In this paper, we propose IMA-GNN as an In-Memory Accelerator for centralized and decentralized Graph Neural Network inference, explore its potential in both settings and provide a guideline for the community targeting flexible and efficient edge computation. Leveraging IMA-GNN, we first model the computation and communication latencies of edge devices. We then present practical case studies on GNN-based taxi demand and supply prediction and also adopt four large graph datasets to quantitatively compare and analyze centralized and decentralized settings. Our cross-layer simulation results demonstrate that on average, IMA-GNN in the centralized setting can obtain ~790x communication speed-up compared to the decentralized GNN setting. However, the decentralized setting performs computation ~1400x faster while reducing the power consumption per device. This further underlines the need for a hybrid semi-decentralized GNN approach.
翻译:在本文中,我们提出了IMA-GNN作为中心化和去中心化图神经网络推断的内存加速器,在两种情况下探索了其潜力,并为社区提供了一个指南,以便针对灵活和高效的边缘计算。借助IMA-GNN,我们首先对边缘设备的计算和通信延迟建模。然后,我们展示了基于GNN的出租车需求和供应预测的实际案例,并采用了四个大型图形数据集,以定量比较和分析中心化和去中心化设置。我们的跨层模拟结果表明,在中心化设置中,IMA-GNN平均可以获得与去中心化GNN设置相比约790倍的通信加速。然而,去中心化设置的计算速度提高了约1400倍,同时降低了每个设备的功耗。这进一步强调了半去中心化GNN方法的需求。