Graph neural networks (GNNs) have demonstrated strong performance on a wide variety of tasks due to their ability to model non-uniform structured data. Despite their promise, there exists little research exploring methods to make them more efficient at inference time. In this work, we explore the viability of training quantized GNNs, enabling the usage of low precision integer arithmetic during inference. We identify the sources of error that uniquely arise when attempting to quantize GNNs, and propose an architecturally-agnostic method, Degree-Quant, to improve performance over existing quantization-aware training baselines commonly used on other architectures, such as CNNs. We validate our method on six datasets and show, unlike previous attempts, that models generalize to unseen graphs. Models trained with Degree-Quant for INT8 quantization perform as well as FP32 models in most cases; for INT4 models, we obtain up to 26% gains over the baselines. Our work enables up to 4.7x speedups on CPU when using INT8 arithmetic.
翻译:图形神经网络(GNNs)由于有能力模拟非统一结构化数据,在各种各样的任务上表现出很强的表现。尽管它们有希望,但几乎没有什么研究探索方法,以使它们在推断时间更有效率。在这项工作中,我们探索了培训量化的GNNs的可行性,以便能够在推断期间使用低精度整数算数。我们找出了在试图对GNS进行量化时唯一出现的错误来源,并提出了建筑-不可知性方法(度-Quant),以改善现有在CNN等其他结构中常用的量化-觉化培训基线的性能。我们验证了我们关于六套数据集的方法,并展示了与以往尝试不同的是,这些模型一般为隐形图形。在为INT8量化而培训的模型和大多数情况下的FD32模型运行;在INT4模型中,我们获得了最多26%的基线收益。我们的工作使得在使用INT8计算时能够对CPU进行4.7的加速。