论文标题
对层次图的对抗攻击神经网络
Adversarial Attack on Hierarchical Graph Pooling Neural Networks
论文作者
论文摘要
近年来见证了图神经网络(GNN)的出现和开发,它们被视为在许多任务中(例如节点分类和图形分类)中图形表示学习的强大方法。这些模型的鲁棒性的研究也开始吸引机器学习领域的关注。但是,该领域的大多数现有工作都集中在节点级任务的GNNS上,而几乎没有完成研究GNNS在图形分类任务中的鲁棒性的工作。在本文中,我们旨在探讨层次图池(HGP)神经网络的脆弱性,它们是高级GNN,在图表分类中表现出色,从预测准确性方面。我们为此任务提出了一个对抗性攻击框架。具体而言,我们设计了一个由卷积和合并操作员组成的代理模型,以生成对抗样本来欺骗基于分层GNN的图形分类模型。我们将保留的节点设置为池操作员作为攻击目标,然后稍微扰动攻击目标,以欺骗层次GNNS中的池操作员,以便他们选择错误的节点来保存。我们显示,我们的替代模型从多个数据集生成的对抗样本具有足够的可传递性来攻击当前的最新图形分类模型。此外,我们在目标模型上进行了强大的火车,并证明了再训练图分类模型能够更好地防御对抗样本的攻击。据我们所知,这是针对基于层次GNN的图形分类模型的对抗性攻击的第一部作品。
Recent years have witnessed the emergence and development of graph neural networks (GNNs), which have been shown as a powerful approach for graph representation learning in many tasks, such as node classification and graph classification. The research on the robustness of these models has also started to attract attentions in the machine learning field. However, most of the existing work in this area focus on the GNNs for node-level tasks, while little work has been done to study the robustness of the GNNs for the graph classification task. In this paper, we aim to explore the vulnerability of the Hierarchical Graph Pooling (HGP) Neural Networks, which are advanced GNNs that perform very well in the graph classification in terms of prediction accuracy. We propose an adversarial attack framework for this task. Specifically, we design a surrogate model that consists of convolutional and pooling operators to generate adversarial samples to fool the hierarchical GNN-based graph classification models. We set the preserved nodes by the pooling operator as our attack targets, and then we perturb the attack targets slightly to fool the pooling operator in hierarchical GNNs so that they will select the wrong nodes to preserve. We show the adversarial samples generated from multiple datasets by our surrogate model have enough transferability to attack current state-of-art graph classification models. Furthermore, we conduct the robust train on the target models and demonstrate that the retrained graph classification models are able to better defend against the attack from the adversarial samples. To the best of our knowledge, this is the first work on the adversarial attack against hierarchical GNN-based graph classification models.