论文标题

反对称DGN:深图网络的稳定体系结构

Anti-Symmetric DGN: a stable architecture for Deep Graph Networks

论文作者

Gravina, Alessio, Bacciu, Davide, Gallicchio, Claudio

论文摘要

深图网络(DGNS)目前在图形学习的研究环境中占主导地位,因为它们的效率和在节点之间实施自适应消息的方案的能力。但是,DGN通常在传播和保留节点之间的长期依赖性的能力上受到限制,即它们患有过度的现象。这降低了它们的有效性,因为预测问题可能需要在不同的(可能是大)的半径上捕获相互作用才能有效解决。在这项工作中,我们提出了反对称的深图网络(A-DGNS),这是一个通过普通微分方程的镜头构想的稳定和非脉冲DGN设计的框架。我们提供了理论上的证据,表明我们的方法是稳定且非疾病的,导致了两个关键结果:保留节点之间的远程信息,并且在培训中没有梯度消失或爆炸发生。我们从经验上验证了几个图基准的拟议方法,表明A-DGN的产量可以提高性能,即使使用了数十层,也可以有效学习。

Deep Graph Networks (DGNs) currently dominate the research landscape of learning from graphs, due to their efficiency and ability to implement an adaptive message-passing scheme between the nodes. However, DGNs are typically limited in their ability to propagate and preserve long-term dependencies between nodes, i.e., they suffer from the over-squashing phenomena. This reduces their effectiveness, since predictive problems may require to capture interactions at different, and possibly large, radii in order to be effectively solved. In this work, we present Anti-Symmetric Deep Graph Networks (A-DGNs), a framework for stable and non-dissipative DGN design, conceived through the lens of ordinary differential equations. We give theoretical proof that our method is stable and non-dissipative, leading to two key results: long-range information between nodes is preserved, and no gradient vanishing or explosion occurs in training. We empirically validate the proposed approach on several graph benchmarks, showing that A-DGN yields to improved performance and enables to learn effectively even when dozens of layers are used.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源