Adversarial training (AT) with samples generated by Fast Gradient Sign Method (FGSM), also known as FGSM-AT, is a computationally simple method to train robust networks. However, during its training procedure, an unstable mode of "catastrophic overfitting" has been identified in arXiv:2001.03994 [cs.LG], where the robust accuracy abruptly drops to zero within a single training step. Existing methods use gradient regularizers or random initialization tricks to attenuate this issue, whereas they either take high computational cost or lead to lower robust accuracy. In this work, we provide the first study, which thoroughly examines a collection of tricks from three perspectives: Data Initialization, Network Structure, and Optimization, to overcome the catastrophic overfitting in FGSM-AT. Surprisingly, we find that simple tricks, i.e., a) masking partial pixels (even without randomness), b) setting a large convolution stride and smooth activation functions, or c) regularizing the weights of the first convolutional layer, can effectively tackle the overfitting issue. Extensive results on a range of network architectures validate the effectiveness of each proposed trick, and the combinations of tricks are also investigated. For example, trained with PreActResNet-18 on CIFAR-10, our method attains 49.8% accuracy against PGD-50 attacker and 46.4% accuracy against AutoAttack, demonstrating that pure FGSM-AT is capable of enabling robust learners. The code and models are publicly available at https://github.com/UCSC-VLAA/Bag-of-Tricks-for-FGSM-AT.
翻译:快速加速信号方法(简称FGSM-AT)生成样本的Aversarial Adversari 培训(AT)是一种计算简单的方法,用于培训强大的网络。然而,在培训过程中,在Arxiv:2001.03994 [cs.LG]中发现了一种不稳定的“灾难性过度”模式,其强力精度在一个培训步骤中突然下降至零。现有方法使用梯度调试器或随机初始化技巧来缓解这一问题,而它们要么采用高计算成本,要么导致更稳健的准确性。在这项工作中,我们提供了第一份研究,从三个角度彻底审查了各种技巧的集合:数据初始化、网络结构和最佳化,以克服FGSM-AT的灾难性过度装配。 令人惊讶的是,我们发现这种简单的把部分像素遮掩蔽(即使没有随机性), b) 设置一个大型的直流化码和平稳启动功能,或者(c) 将第一个精准的精准性ODAL-AT-RODA的精度定度放在第一个变精度模型上。