论文标题
图形神经网络可学习的可交换性单体
Learnable Commutative Monoids for Graph Neural Networks
论文作者
论文摘要
图形神经网络(GNN)已被证明对聚集功能的选择高度敏感。 Cohen-Karlik等人可以将节点的邻居求和可以近似于离散输入的任何置换函数。 [2020]证明存在一些集合构型问题,而这些问题不能概括为无限的输入,提出了针对置换不变性作为更具表现力的聚合器的经常性神经网络。我们表明,这些结果延续到了图形域:配备了复发器聚合器的GNN在合成基准和现实世界中的问题上都与最新的置换式聚合器具有竞争力。但是,尽管复发器的好处有好处,但它们的$ O(v)$深度使它们既难以并行,又难以在大图上进行训练。受到观察的启发,即GNN行为良好的聚合器是其潜在空间的交换性单体,我们提出了一个框架,用于构建可学习的,交换性的关联,关联的二进制运算符。因此,我们构建了一个$ o(\ log V)$深度的聚合器,为并行性和依赖性长度带来了指数的改进,同时与经常性聚合器竞争性能竞争。基于我们的经验观察,我们提出的可学习的交换性单体(LCM)聚合器代表了有效的和表达的聚合器之间的良好权衡。
Graph neural networks (GNNs) have been shown to be highly sensitive to the choice of aggregation function. While summing over a node's neighbours can approximate any permutation-invariant function over discrete inputs, Cohen-Karlik et al. [2020] proved there are set-aggregation problems for which summing cannot generalise to unbounded inputs, proposing recurrent neural networks regularised towards permutation-invariance as a more expressive aggregator. We show that these results carry over to the graph domain: GNNs equipped with recurrent aggregators are competitive with state-of-the-art permutation-invariant aggregators, on both synthetic benchmarks and real-world problems. However, despite the benefits of recurrent aggregators, their $O(V)$ depth makes them both difficult to parallelise and harder to train on large graphs. Inspired by the observation that a well-behaved aggregator for a GNN is a commutative monoid over its latent space, we propose a framework for constructing learnable, commutative, associative binary operators. And with this, we construct an aggregator of $O(\log V)$ depth, yielding exponential improvements for both parallelism and dependency length while achieving performance competitive with recurrent aggregators. Based on our empirical observations, our proposed learnable commutative monoid (LCM) aggregator represents a favourable tradeoff between efficient and expressive aggregators.