Recent years have witnessed the deployment of adversarial attacks to evaluate the robustness of Neural Networks. Past work in this field has relied on traditional optimization algorithms that ignore the inherent structure of the problem and data, or generative methods that rely purely on learning and often fail to generate adversarial examples where they are hard to find. To alleviate these deficiencies, we propose a novel attack based on a graph neural network (GNN) that takes advantage of the strengths of both approaches; we call it AdvGNN. Our GNN architecture closely resembles the network we wish to attack. During inference, we perform forward-backward passes through the GNN layers to guide an iterative procedure towards adversarial examples. During training, its parameters are estimated via a loss function that encourages the efficient computation of adversarial examples over a time horizon. We show that our method beats state-of-the-art adversarial attacks, including PGD-attack, MI-FGSM, and Carlini and Wagner attack, reducing the time required to generate adversarial examples with small perturbation norms by over 65\%. Moreover, AdvGNN achieves good generalization performance on unseen networks. Finally, we provide a new challenging dataset specifically designed to allow for a more illustrative comparison of adversarial attacks.
翻译:近些年来,我们目睹了对立攻击的部署,以评价神经网络的强健性。过去在这一领域的工作依赖传统的优化算法,这种算法忽视了问题和数据的内在结构,或纯粹依赖学习的基因化方法,往往没有产生难以找到的对抗性例子。为了减轻这些缺陷,我们提议以图表神经网络(GNN)为基础,进行新的攻击,利用两种方法的优势;我们称之为AdvGNNN。我们的GNN结构与我们希望攻击的网络非常相似。在推论期间,我们通过GNN层向前走过后路,以引导一种迭接程序来树立对抗性例子。在培训期间,通过一种鼓励在时间跨度上有效计算对抗性例子的损失函数来估计其参数。我们表明,我们的方法战胜了最先进的对立性攻击,包括PGD-攻击、MI-FGMSM、Carlini和Wagner攻击,从而缩短了产生具有小的对立性攻击规范所需的时间。此外,AdvGNNNN能够具体地对新的对抗性攻击进行具有挑战性的数据比较。最后,我们为对抗性攻击提供了对立性攻击提供了对立性攻击的比较。