Deep Graph Networks (DGNs) currently dominate the research landscape of learning from graphs, due to their efficiency and ability to implement an adaptive message-passing scheme between the nodes. However, DGNs are typically limited in their ability to propagate and preserve long-term dependencies between nodes, \ie they suffer from the over-squashing phenomena. This reduces their effectiveness, since predictive problems may require to capture interactions at different, and possibly large, radii in order to be effectively solved. In this work, we present Anti-Symmetric Deep Graph Networks (A-DGNs), a framework for stable and non-dissipative DGN design, conceived through the lens of ordinary differential equations. We give theoretical proof that our method is stable and non-dissipative, leading to two key results: long-range information between nodes is preserved, and no gradient vanishing or explosion occurs in training. We empirically validate the proposed approach on several graph benchmarks, showing that A-DGN yields to improved performance and enables to learn effectively even when dozens of layers are used.
翻译:深图网络(DGNs)目前支配着从图表中学习的研究领域,这是因为它们的效率和执行节点之间适应性信息传递计划的能力。然而,DGNs通常在传播和维护节点之间长期依赖性的能力方面受到限制,它们遭受着过度隔绝现象的影响。这降低了其有效性,因为预测问题可能需要在不同、甚至可能很大的地方捕捉相互作用,才能有效解决。在这项工作中,我们提出了反对称深度图网络(A-DGNs),这是稳定和非分散性DGN设计的框架,通过普通差异方程式的透镜构思。我们从理论上证明,我们的方法是稳定的、非分散的,导致两个主要结果:节点之间的远程信息得以保留,在培训中不会发生梯度消失或爆炸。我们从经验上验证了几个图表基准的拟议方法,表明A-DGN的产量可以提高性能,即使使用了几十层,也能有效地学习。