A cursory reading of the literature suggests that we have made a lot of progress in designing effective adversarial defenses for Graph Neural Networks (GNNs). Yet, the standard methodology has a serious flaw - virtually all of the defenses are evaluated against non-adaptive attacks leading to overly optimistic robustness estimates. We perform a thorough robustness analysis of 7 of the most popular defenses spanning the entire spectrum of strategies, i.e., aimed at improving the graph, the architecture, or the training. The results are sobering - most defenses show no or only marginal improvement compared to an undefended baseline. We advocate using custom adaptive attacks as a gold standard and we outline the lessons we learned from successfully designing such attacks. Moreover, our diverse collection of perturbed graphs forms a (black-box) unit test offering a first glance at a model's robustness.
翻译:对文献的粗略阅读表明,我们在设计图神经网络(GNNs)的有效对抗性防御方面取得了很大进展。 然而,标准方法有一个严重的缺陷 — — 几乎所有防御都针对非适应性攻击进行了评估,导致过于乐观的稳健性估计。 我们对7个最受欢迎的防御系统进行了彻底的稳健性分析,这些防御系统覆盖了整个战略的方方面面,即旨在改进图表、结构或培训。结果正在清醒中 — — 多数防御系统与未设防基线相比,没有或只是略有改善。我们提倡使用习惯性适应性攻击作为金本位,我们概述了我们从成功设计此类攻击中吸取的经验教训。此外,我们收集的周遭图形也形成了一个(黑盒)单元测试,首次审视模型的坚固性。