Recently, graph convolutional networks (GCNs) have shown to be vulnerable to small adversarial perturbations, which becomes a severe threat and largely limits their applications in security-critical scenarios. To mitigate such a threat, considerable research efforts have been devoted to increasing the robustness of GCNs against adversarial attacks. However, current approaches for defense are typically designed for the whole graph and consider the global performance, posing challenges in protecting important local nodes from stronger adversarial targeted attacks. In this work, we present a simple yet effective method, named \textbf{\underline{G}}raph \textbf{\underline{U}}niversal \textbf{\underline{A}}dve\textbf{\underline{R}}sarial \textbf{\underline{D}}efense (GUARD). Unlike previous works, GUARD protects each individual node from attacks with a universal defensive patch, which is generated once and can be applied to any node (node-agnostic) in a graph. Extensive experiments on four benchmark datasets demonstrate that our method significantly improves robustness for several established GCNs against multiple adversarial attacks and outperforms existing adversarial defense methods by large margins. Our code is publicly available at https://github.com/EdisonLeeeee/GUARD.
翻译:最近,图形革命网络(GCNs)显示很容易受到小型对抗性攻击,这种攻击成为严重威胁,并在很大程度上限制了其在安全危急情况下的应用。为了减轻这种威胁,已经进行了大量研究,以提高GCN对对抗性攻击的稳健性。然而,目前的防御方法一般是为整张图设计的,并考虑全球表现,对保护重要的当地节点免受更强烈的对抗性攻击构成挑战。在这项工作中,我们提出了一个简单而有效的方法,名为\textbfsunderline{G ⁇ raph\ textbf_underline{U ⁇ niversal\ textbf_underline{A ⁇ dve\textbf_underline{Räunderline{R ⁇ saarial\tbf_debuf=underline{Duderline{GARDCDRD) 。与以前的工作不同,GURD(GUARD)保护每个个人节点不受普遍防御性攻击,它曾经产生,并且可以应用在图表中的任何节点(nde-nostic) 。在四种基准数据基点上进行广泛的实验,对四种基准数据显示,我们对抗性防御性攻击的方法大大改进了我们现有的防御性准则。在多种对抗性/对抗性攻击上已经建立起来/防御性准则。