论文标题
重新访问对图神经网络的对抗性攻击以进行图形分类
Revisiting Adversarial Attacks on Graph Neural Networks for Graph Classification
论文作者
论文摘要
图形神经网络(GNN)在图形分类及其多样化的下游现实世界应用方面取得了巨大的成功。尽管学习图表取得了巨大成功,但当前的GNN模型已经证明了它们在图形结构数据上可能存在的对抗性示例的脆弱性。现有的方法要么仅限于结构攻击,要么仅限于本地信息,敦促在图形分类上设计一个更一般的攻击框架,由于使用全球图形信息产生本地节点级的对抗示例,因此面临重大挑战。为了解决这一“全球到本地”攻击挑战,我们提出了一个新颖的一般框架,以通过操纵图结构和节点特征来生成对抗性示例。具体而言,我们利用图类别激活映射及其变体来产生与图形分类任务相对应的节点级的重要性。然后,通过算法的启发式设计,我们可以在节点级别和子图级重要性的帮助下在不明显的扰动预算下执行功能和结构攻击。在六个现实世界基准上攻击四个最先进的图形分类模型的实验验证了我们框架的灵活性和有效性。
Graph neural networks (GNNs) have achieved tremendous success in the task of graph classification and its diverse downstream real-world applications. Despite the huge success in learning graph representations, current GNN models have demonstrated their vulnerability to potentially existent adversarial examples on graph-structured data. Existing approaches are either limited to structure attacks or restricted to local information, urging for the design of a more general attack framework on graph classification, which faces significant challenges due to the complexity of generating local-node-level adversarial examples using the global-graph-level information. To address this "global-to-local" attack challenge, we present a novel and general framework to generate adversarial examples via manipulating graph structure and node features. Specifically, we make use of Graph Class Activation Mapping and its variant to produce node-level importance corresponding to the graph classification task. Then through a heuristic design of algorithms, we can perform both feature and structure attacks under unnoticeable perturbation budgets with the help of both node-level and subgraph-level importance. Experiments towards attacking four state-of-the-art graph classification models on six real-world benchmarks verify the flexibility and effectiveness of our framework.