Oversmoothing is a central challenge of building more powerful Graph Neural Networks (GNNs). While previous works have only demonstrated that oversmoothing is inevitable when the number of graph convolutions tends to infinity, in this paper, we precisely characterize the mechanism behind the phenomenon via a non-asymptotic analysis. Specifically, we distinguish between two different effects when applying graph convolutions -- an undesirable mixing effect that homogenizes node representations in different classes, and a desirable denoising effect that homogenizes node representations in the same class. By quantifying these two effects on random graphs sampled from the Contextual Stochastic Block Model (CSBM), we show that oversmoothing happens once the mixing effect starts to dominate the denoising effect, and the number of layers required for this transition is $O(\log N/\log (\log N))$ for sufficiently dense graphs with $N$ nodes. We also extend our analysis to study the effects of Personalized PageRank (PPR), or equivalently, the effects of initial residual connections on oversmoothing. Our results suggest that while PPR mitigates oversmoothing at deeper layers, PPR-based architectures still achieve their best performance at a shallow depth and are outperformed by the graph convolution approach on certain graphs. Finally, we support our theoretical results with numerical experiments, which further suggest that the oversmoothing phenomenon observed in practice can be magnified by the difficulty of optimizing deep GNN models.
翻译:过度透析是建设更强大的图形神经网络(GNNS)的一个中心挑战。 虽然先前的著作仅表明,当图形变形的数量趋向于无限时,过度透析是不可避免的, 但在本文件中,我们通过非无损分析准确地区分了该现象背后的机制。 具体地说,我们在应用图形变形时区分了两种不同的效果 -- -- 一种不可取的混合效应,在不同类别中将结点表示同质化,以及一种理想的分解效应,使同一类别中的节点表示趋于一致。通过量化这两个对从上下文的深层颗粒模型(CSBM)取样的随机图形的影响,我们表明,一旦混合效应开始支配消化效应,就会出现超浮现象,而这一转变所需的层数是$O (\log N/\log (\N)),对于以美元计数的足够稠密的图形,我们还将我们的分析扩大到研究个人化的PeRCK(PPR)的影响,或者同样地,通过我们观察到的深度变相偏化的图形分析模型来减轻其初步残余连接的影响,而我们最终的图像变色结构则显示, 最深层结果可以显示,我们最深层的更深层的更深层分析结果。我们的结果可以显示,我们最终的平流的平局将显示为最深层的更深层的平。</s>