论文标题
代表图形神经网络的远程上下文,全球关注
Representing Long-Range Context for Graph Neural Networks with Global Attention
论文作者
论文摘要
图神经网络是结构化数据集的强大体系结构。但是,当前的方法难以代表远程依赖性。扩展GNN的深度或宽度不足以扩大接受场,因为较大的GNN遇到优化的不稳定性,例如消失的梯度和表示过度厚度,而基于汇总的方法尚未在计算机视觉中具有普遍有用。在这项工作中,我们建议使用基于变压器的自我注意力来学习远程成对关系,并采用一种新颖的“读数”机制来获得全局图嵌入。受到最新计算机视觉结果的启发,在学习长期关系中,我们称为GraphTrans,在学习长期关系中发现了位置不变的注意力,它在标准GNN模块之后应用了置换不变的变压器模块。这种简单的体系结构导致了几个图形分类任务的最新结果,超过明确编码图形结构的方法。我们的结果表明,没有图形结构的纯学习方法可能适合在图上学习高级,远程关系。 GraphTrans的代码可从https://github.com/ucbrise/graphtrans获得。
Graph neural networks are powerful architectures for structured datasets. However, current methods struggle to represent long-range dependencies. Scaling the depth or width of GNNs is insufficient to broaden receptive fields as larger GNNs encounter optimization instabilities such as vanishing gradients and representation oversmoothing, while pooling-based approaches have yet to become as universally useful as in computer vision. In this work, we propose the use of Transformer-based self-attention to learn long-range pairwise relationships, with a novel "readout" mechanism to obtain a global graph embedding. Inspired by recent computer vision results that find position-invariant attention performant in learning long-range relationships, our method, which we call GraphTrans, applies a permutation-invariant Transformer module after a standard GNN module. This simple architecture leads to state-of-the-art results on several graph classification tasks, outperforming methods that explicitly encode graph structure. Our results suggest that purely-learning-based approaches without graph structure may be suitable for learning high-level, long-range relationships on graphs. Code for GraphTrans is available at https://github.com/ucbrise/graphtrans.