Deep Graph Networks (DGNs) currently dominate the research landscape of learning from graphs, due to their efficiency and ability to implement an adaptive message-passing scheme between the nodes. However, DGNs are typically limited in their ability to propagate and preserve long-term dependencies between nodes, i.e., they suffer from the over-squashing phenomena. This reduces their effectiveness, since predictive problems may require to capture interactions at different, and possibly large, radii in order to be effectively solved. In this work, we present Anti-Symmetric Deep Graph Networks (A-DGNs), a framework for stable and non-dissipative DGN design, conceived through the lens of ordinary differential equations. We give theoretical proof that our method is stable and non-dissipative, leading to two key results: long-range information between nodes is preserved, and no gradient vanishing or explosion occurs in training. We empirically validate the proposed approach on several graph benchmarks, showing that A-DGN yields to improved performance and enables to learn effectively even when dozens of layers are used.
翻译:深图网络(DGNs)目前支配着从图表中学习的研究领域,这是因为它们的效率和能力,可以在节点之间执行适应性的信息传递计划。然而,DGNs通常在传播和维护节点之间长期依赖性的能力方面受到限制,即它们受到过度隔绝现象的影响。这降低了它们的效力,因为预测问题可能需要在不同的、甚至大半射线上捕捉相互作用,才能有效解决。在这项工作中,我们介绍了反对称深度图网络(A-DGNs),这是稳定和非分散的DGN设计框架,通过普通差异方程式的透镜构思。我们从理论上证明我们的方法是稳定的、非分散的,导致两个主要结果:节点之间的长程信息得到保存,培训中不会发生梯度消失或爆炸。我们从经验上验证了几个图表基准的拟议方法,表明A-DGN的产量可以提高性能,即使使用了几十层,也能有效地学习。