论文标题

图形通用对抗性攻击:一些不好的演员破坏图形学习模型

Graph Universal Adversarial Attacks: A Few Bad Actors Ruin Graph Learning Models

论文作者

Zang, Xiao, Xie, Yi, Chen, Jie, Yuan, Bo

论文摘要

深度神经网络虽然良好,但已知对小对抗扰动敏感。这种现象构成了严重的安全威胁,并要求对深度学习模型的鲁棒性进行深入研究。随着用于图结构数据的神经网络的出现,敦促类似的研究以了解其稳健性。已经发现,对抗图形结构和/或节点特征的对抗可能会导致模型性能的显着降解。在这项工作中,我们从不同的角度表明,如果该图包含一些坏角色节点,这种脆弱性也会同样发生,这些节点通过将连接转换为任何有针对性的受害者,从而损害了受过训练的图形神经网络。更糟糕的是,发现一个图形模型的不良演员也严重损害了其他模型。我们称不良演员``锚节点'',并提出了一种名为GUA的算法来识别它们。彻底的实证研究表明,一个有趣的发现,即锚节点通常属于同一类。他们还证实了锚节点数量与攻击成功率之间的直观权衡。对于包含2708个节点的数据集CORA,只有六个锚节点的六个锚节点将导致GCN和其他三个模型的攻击成功率高于80 \%。

Deep neural networks, while generalize well, are known to be sensitive to small adversarial perturbations. This phenomenon poses severe security threat and calls for in-depth investigation of the robustness of deep learning models. With the emergence of neural networks for graph structured data, similar investigations are urged to understand their robustness. It has been found that adversarially perturbing the graph structure and/or node features may result in a significant degradation of the model performance. In this work, we show from a different angle that such fragility similarly occurs if the graph contains a few bad-actor nodes, which compromise a trained graph neural network through flipping the connections to any targeted victim. Worse, the bad actors found for one graph model severely compromise other models as well. We call the bad actors ``anchor nodes'' and propose an algorithm, named GUA, to identify them. Thorough empirical investigations suggest an interesting finding that the anchor nodes often belong to the same class; and they also corroborate the intuitive trade-off between the number of anchor nodes and the attack success rate. For the dataset Cora which contains 2708 nodes, as few as six anchor nodes will result in an attack success rate higher than 80\% for GCN and other three models.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源