Yang Runze, Long Teng
School of Information Engineering, China University of Geosciences, Beijing, China.
PeerJ Comput Sci. 2021 Aug 24;7:e693. doi: 10.7717/peerj-cs.693. eCollection 2021.
In recent years, graph convolutional networks (GCNs) have emerged rapidly due to their excellent performance in graph data processing. However, recent researches show that GCNs are vulnerable to adversarial attacks. An attacker can maliciously modify edges or nodes of the graph to mislead the model's classification of the target nodes, or even cause a degradation of the model's overall classification performance. In this paper, we first propose a black-box adversarial attack framework based on derivative-free optimization (DFO) to generate graph adversarial examples without using gradient and apply advanced DFO algorithms conveniently. Second, we implement a direct attack algorithm (DFDA) using the Nevergrad library based on the framework. Additionally, we overcome the problem of large search space by redesigning the perturbation vector using constraint size. Finally, we conducted a series of experiments on different datasets and parameters. The results show that DFDA outperforms Nettack in most cases, and it can achieve an average attack success rate of more than 95% on the Cora dataset when perturbing at most eight edges. This demonstrates that our framework can fully exploit the potential of DFO methods in node classification adversarial attacks.
近年来,图卷积网络(GCN)因其在图数据处理方面的卓越性能而迅速兴起。然而,最近的研究表明,GCN容易受到对抗性攻击。攻击者可以恶意修改图的边或节点,以误导模型对目标节点的分类,甚至导致模型整体分类性能下降。在本文中,我们首先提出了一种基于无导数优化(DFO)的黑盒对抗性攻击框架,以在不使用梯度的情况下生成图对抗性示例,并方便地应用先进的DFO算法。其次,我们基于该框架使用Nevergrad库实现了一种直接攻击算法(DFDA)。此外,我们通过使用约束大小重新设计扰动向量来克服搜索空间过大的问题。最后,我们在不同的数据集和参数上进行了一系列实验。结果表明,DFDA在大多数情况下优于Nettack,并且在Cora数据集上最多扰动八条边时,平均攻击成功率可以达到95%以上。这表明我们的框架可以充分发挥DFO方法在节点分类对抗性攻击中的潜力。