Graph neural networks (GNN) suffer from severe inefficiency. It is mainly caused by the exponential growth of node dependency with the increase of layers. It extremely limits the application of stochastic optimization algorithms so that the training of GNN is usually time-consuming. To address this problem, we propose to decouple a multi-layer GNN as multiple simple modules for more efficient training, which is comprised of classical forward training (FT)and designed backward training (BT). Under the proposed framework, each module can be trained efficiently in FT by stochastic algorithms without distortion of graph information owing to its simplicity. To avoid the only unidirectional information delivery of FT and sufficiently train shallow modules with the deeper ones, we develop a backward training mechanism that makes the former modules perceive the latter modules. The backward training introduces the reversed information delivery into the decoupled modules as well as the forward information delivery. To investigate how the decoupling and greedy training affect the representational capacity, we theoretically prove that the error produced by linear modules will not accumulate on unsupervised tasks in most cases. The theoretical and experimental results show that the proposed framework is highly efficient with reasonable performance.
翻译:图神经网络(GNN)具有严重的效率问题,主要是由于随着层数的增加,节点依赖性呈指数增长。这导致了随机优化算法的应用受限,使得GNN的训练通常很耗时。为了解决这个问题,我们提出了将多层GNN分解为多个简单模块,以进行更有效的训练,其中包括传统的正向训练(FT)和设计的反向训练(BT)。在所提出的框架下,每个模块都可以通过随机算法在FT中有效地进行训练,由于其简单性,不会扭曲图形信息。为了避免FT的单向信息传递并且足够地训练较浅的模块与更深的模块,我们开发了一个反向训练机制,使前面的模块感知后面的模块。反向训练引入了反向信息传递以及前向信息传递到分离的模块中。为了研究分解和贪婪训练如何影响表现能力,我们在理论上证明线性模块产生的误差在大多数情况下不会在无监督任务中累积。理论和实验结果表明,所提出的框架具有高效性和合理的性能。