Graph convolutional networks (GCNs) and their variants have achieved great success in dealing with graph-structured data. However, it is well known that deep GCNs will suffer from over-smoothing problem, where node representations tend to be indistinguishable as we stack up more layers. Although extensive research has confirmed this prevailing understanding, few theoretical analyses have been conducted to study the expressivity and trainability of deep GCNs. In this work, we demonstrate these characterizations by studying the Gaussian Process Kernel (GPK) and Graph Neural Tangent Kernel (GNTK) of an infinitely-wide GCN, corresponding to the analysis on expressivity and trainability, respectively. We first prove the expressivity of infinitely-wide GCNs decaying at an exponential rate by applying the mean-field theory on GPK. Besides, we formulate the asymptotic behaviors of GNTK in the large depth, which enables us to reveal the dropping trainability of wide and deep GCNs at an exponential rate. Additionally, we extend our theoretical framework to analyze residual connection-resemble techniques. We found that these techniques can mildly mitigate exponential decay, but they failed to overcome it fundamentally. Finally, all theoretical results in this work are corroborated experimentally on a variety of graph-structured datasets.
翻译:虽然广泛的研究证实了这种普遍的理解,但很少进行理论分析,以研究深层GCN的表达性和可训练性。在这项工作中,我们通过研究高斯进程核心内尔(GPK)和深层GNCN(GNTK)的无限范围GCN(GNTK)的可训练性来展示这些特征。此外,我们首先通过在GPK上应用中位理论来证明无限范围的GCN以指数速度衰落的显性。此外,我们把理论框架扩展到分析离心力和可训练性。此外,我们把这些理论框架扩大到分析离心力和深层GCN(GNTK)的可训练性下降。我们发现,它们可以以指数速度显示广度和深层GNCN(GNTKK)的可训练性下降性。此外,我们把理论框架扩大到分析离心力和可训练力的离心力的理论结构。我们把这种理论框架扩大到了最深层的实验性研究,最终,我们发现这些理论基础的实验性实验技术可以减轻了。