Adversarial training is exploited to develop a robust Deep Neural Network (DNN) model against the malicious altered data. These attacks may have catastrophic effects on DNN models but are indistinguishable for a human being. For example, an external attack can modify an image adding noises invisible for a human eye, but a DNN model misclassified the image. A key objective for developing robust DNN models is to use a learning algorithm that is fast but can also give model that is robust against different types of adversarial attacks. Especially for adversarial training, enormously long training times are needed for obtaining high accuracy under many different types of adversarial samples generated using different adversarial attack techniques. This paper aims at accelerating the adversarial training to enable fast development of robust DNN models against adversarial attacks. The general method for improving the training performance is the hyperparameters fine-tuning, where the learning rate is one of the most crucial hyperparameters. By modifying its shape (the value over time) and value during the training, we can obtain a model robust to adversarial attacks faster than standard training. First, we conduct experiments on two different datasets (CIFAR10, CIFAR100), exploring various techniques. Then, this analysis is leveraged to develop a novel fast training methodology, AccelAT, which automatically adjusts the learning rate for different epochs based on the accuracy gradient. The experiments show comparable results with the related works, and in several experiments, the adversarial training of DNNs using our AccelAT framework is conducted up to 2 times faster than the existing techniques. Thus, our findings boost the speed of adversarial training in an era in which security and performance are fundamental optimization objectives in DNN-based applications.
翻译:利用Adversari 培训开发一个强大的深神经网络模型以对抗恶意改变数据。这些攻击可能给DNN模型带来灾难性后果,但对人类来说无法区分。例如,外部攻击可以改变增加人类眼睛看不见的噪音的图像,但DNN模型对图像进行错误分类。开发强大的DNN模型的一个关键目标是使用一种快速的学习算法,但也能够给模型提供对不同类型对抗性攻击的强力。特别是对于对抗性训练来说,在使用不同的对抗性攻击技术生成的许多不同类型的对抗性样本中,要达到高精确度,则需要很长的培训时间。例如,外部攻击可以改变一个为人类眼睛看不见的噪音添加噪音的图像,但DNNNNN模型则需要非常长的培训时间。例如,为了加速对敌对性攻击的快速发展强大的DNNNNN模型模型。 提高培训性能的一般方法是超参数的微调,其中学习率是最重要的超标值。 通过修改其形状(基于时间的价值)和价值,我们可以得到比标准训练更快的模型。首先,我们用两种可比较的训练方法对两种不同的训练方法进行实验, 进行升级的NEARAT标准的实验,在快速分析中进行。