论文标题
Tuneup:用于图形神经网络的简单改进的培训策略
TuneUp: A Simple Improved Training Strategy for Graph Neural Networks
论文作者
论文摘要
尽管图形神经网络(GNN)最近取得了进步,但他们的训练策略在很大程度上仍然不足。传统的培训策略在原始图中同样了解所有节点,这可能是最佳的,因为某些节点通常比其他节点更难学习。在这里,我们提出了Tuneup,这是一种简单的基于课程的培训策略,用于提高GNN的预测性能。 Tuneup在两个阶段训练GNN。在第一阶段,Tuneup应用常规训练以获得强大的基础GNN。基本GNN倾向于在头节点(具有较大程度的节点)上表现良好,但在尾部节点(小度的节点)上的表现较少。因此,Tuneup的第二阶段着重于通过进一步训练合成生成的尾部节点数据的基础GNN来改善对困难尾部节点的预测。我们从理论上分析了Tuneup,并证明它可以提高尾部节点上的泛化性能。 Tuneup易于实现,并且适用于广泛的GNN架构和预测任务。对五种不同的GNN架构,三种预测任务以及转导性和归纳性环境的Tuneup进行了广泛的评估表明,Tuneup显着提高了基本GNN在尾部节点上的性能,同时通常会改善头部节点上的性能。总体而言,Tuneup分别在转导性和挑战性归纳环境中产生高达57.6%和92.2%的相对预测性能提高。
Despite recent advances in Graph Neural Networks (GNNs), their training strategies remain largely under-explored. The conventional training strategy learns over all nodes in the original graph(s) equally, which can be sub-optimal as certain nodes are often more difficult to learn than others. Here we present TuneUp, a simple curriculum-based training strategy for improving the predictive performance of GNNs. TuneUp trains a GNN in two stages. In the first stage, TuneUp applies conventional training to obtain a strong base GNN. The base GNN tends to perform well on head nodes (nodes with large degrees) but less so on tail nodes (nodes with small degrees). Therefore, the second stage of TuneUp focuses on improving prediction on the difficult tail nodes by further training the base GNN on synthetically generated tail node data. We theoretically analyze TuneUp and show it provably improves generalization performance on tail nodes. TuneUp is simple to implement and applicable to a broad range of GNN architectures and prediction tasks. Extensive evaluation of TuneUp on five diverse GNN architectures, three types of prediction tasks, and both transductive and inductive settings shows that TuneUp significantly improves the performance of the base GNN on tail nodes, while often even improving the performance on head nodes. Altogether, TuneUp produces up to 57.6% and 92.2% relative predictive performance improvement in the transductive and the challenging inductive settings, respectively.