Great advances in deep neural networks (DNNs) have led to state-of-the-art performance on a wide range of tasks. However, recent studies have shown that DNNs are vulnerable to adversarial attacks, which have brought great concerns when deploying these models to safety-critical applications such as autonomous driving. Different defense approaches have been proposed against adversarial attacks, including: a) empirical defenses, which can usually be adaptively attacked again without providing robustness certification; and b) certifiably robust approaches, which consist of robustness verification providing the lower bound of robust accuracy against any attacks under certain conditions and corresponding robust training approaches. In this paper, we systematize certifiably robust approaches and related practical and theoretical implications and findings. We also provide the first comprehensive benchmark on existing robustness verification and training approaches on different datasets. In particular, we 1) provide a taxonomy for the robustness verification and training approaches, as well as summarize the methodologies for representative algorithms, 2) reveal the characteristics, strengths, limitations, and fundamental connections among these approaches, 3) discuss current research progresses, theoretical barriers, main challenges, and future directions for certifiably robust approaches for DNNs, and 4) provide an open-sourced unified platform to evaluate 20+ representative certifiably robust approaches.
翻译:深入的神经网络(DNNs)取得了巨大进步,导致在一系列广泛任务上取得了最先进的实绩;然而,最近的研究表明,DNNs很容易受到对抗性攻击,这在将这些模型用于安全关键应用(如自主驱动)时引起了极大的关注;针对对抗性攻击提出了不同的防御办法,包括:(a) 经验防御办法,通常可以在不提供稳健认证的情况下再次进行适应性攻击;和(b) 可证实的稳健办法,包括稳健的核查,在特定条件下对任何袭击提供较弱的稳健准确性约束,以及相应的稳健培训办法;在本文件中,我们系统化了可靠的稳健办法以及相关的实际和理论影响和结论;我们还为现有稳健的稳健性核查和不同数据集的培训办法提供了第一个全面基准;特别是,我们(1) 为稳健健的核查和培训办法提供分类,并概述代表性算法的方法;(2) 揭示这些办法的特点、强、强、局限性和基本联系;(3) 讨论目前的研究进展、理论障碍、主要挑战和未来方向;(4) 可靠地为有代表性的20个可靠、可靠、可靠、可靠、可靠、可靠、可靠、可靠的平台提供可靠的方法。