Single-step adversarial training (AT) has received wide attention as it proved to be both efficient and robust. However, a serious problem of catastrophic overfitting exists, i.e., the robust accuracy against projected gradient descent (PGD) attack suddenly drops to $0\%$ during the training. In this paper, we understand this problem from a novel perspective of optimization and firstly reveal the close link between the fast-growing gradient of each sample and overfitting, which can also be applied to understand the robust overfitting phenomenon in multi-step AT. To control the growth of the gradient during the training, we propose a new AT method, subspace adversarial training (Sub-AT), which constrains the AT in a carefully extracted subspace. It successfully resolves both two kinds of overfitting and hence significantly boosts the robustness. In subspace, we also allow single-step AT with larger steps and larger radius, which further improves the robustness performance. As a result, we achieve the state-of-the-art single-step AT performance: our pure single-step AT can reach over $\mathbf{51}\%$ robust accuracy against strong PGD-50 attack with radius $8/255$ on CIFAR-10, even surpassing the standard multi-step PGD-10 AT with huge computational advantages. The code is released$\footnote{\url{https://github.com/nblt/Sub-AT}}$.
翻译:单步对抗性训练(AT)已经得到广泛关注,因为它证明既有效又有力。然而,一个严重的灾难性过度改造问题仍然存在,即对预测的梯度下降(PGD)攻击的精确度在培训期间突然下降到0美元。在本文中,我们从优化的新角度来理解这一问题,并首先揭示每个样本的快速增长梯度和超装之间的密切联系,这也可用于理解多步制的强力超配现象。为了控制培训期间的梯度增长,我们提出了一个新的AT方法,即亚空间对抗性攻击(Sub-AT),这限制了AT在一个仔细提取的子空间中。它成功地解决了两种过度匹配,从而大大提升了稳健性。在亚空间中,我们还允许以更大的步骤和更大的半径单步制来进一步提高稳健性性表现。结果是,我们实现了在多步制AT的单步制表现:我们纯单步制AT可以超过$mab$/Sub_AT$(Sub-AT)的亚盘训练(Sub-AT),这限制了AT在仔细提取的亚基值的亚基标准/QRGDGD/10的多步制标准。