The performance limit of Graph Convolutional Networks (GCNs) and the fact that we cannot stack more of them to increase the performance, which we usually do for other deep learning paradigms, are pervasively thought to be caused by the limitations of the GCN layers, including insufficient expressive power, etc. However, if so, for a fixed architecture, it would be unlikely to lower the training difficulty and to improve performance by changing only the training procedure, which we show in this paper not only possible but possible in several ways. This paper first identify the training difficulty of GCNs from the perspective of graph signal energy loss. More specifically, we find that the loss of energy in the backward pass during training nullifies the learning of the layers closer to the input. Then, we propose several methodologies to mitigate the training problem by slightly modifying the GCN operator, from the energy perspective. After empirical validation, we confirm that these changes of operator lead to significant decrease in the training difficulties and notable performance boost, without changing the composition of parameters. With these, we conclude that the root cause of the problem is more likely the training difficulty than the others.
翻译:图表革命网络(GCN)的绩效限制,以及我们不能为了提高绩效而把其中更多的内容叠叠起来,而这是我们通常为其他深层学习范式而做的。 人们普遍认为,这是GCN层的局限性造成的,包括表达力不足等等。 然而,如果是这样,对于固定的建筑来说,我们不可能通过仅仅改变培训程序来降低培训难度和改进绩效,我们在本文件中不仅可能,而且以几种方式都表明了这种程序。本文件首先从图表显示的能源损失的角度确定GCN的培训难度。更具体地说,我们发现,在培训过程中落后的能源损失使得学习与投入更接近的层无法学习。然后,我们从能源的角度提出一些方法,通过稍微修改GCN运营商来缓解培训问题。在经验验证后,我们确认操作商的这些变化导致培训困难和显著的绩效提升显著减少,而没有改变参数的构成。我们的结论是,问题的根源是培训困难比其他人更为严重。