Convolutional neural networks have outperformed humans in image recognition tasks, but they remain vulnerable to attacks from adversarial examples. Since these data are crafted by adding imperceptible noise to normal images, their existence poses potential security threats to deep learning systems. Sophisticated adversarial examples with strong attack performance can also be used as a tool to evaluate the robustness of a model. However, the success rate of adversarial attacks can be further improved in black-box environments. Therefore, this study combines a modified Adam gradient descent algorithm with the iterative gradient-based attack method. The proposed Adam Iterative Fast Gradient Method is then used to improve the transferability of adversarial examples. Extensive experiments on ImageNet showed that the proposed method offers a higher attack success rate than existing iterative methods. By extending our method, we achieved a state-of-the-art attack success rate of 95.0% on defense models.
翻译:进化神经网络在图像识别任务方面超过了人类,但是它们仍然容易受到来自对抗性实例的攻击。由于这些数据是通过在正常图像中添加无法察觉的噪音而形成的,因此它们的存在对深层学习系统构成了潜在的安全威胁。具有强力攻击性能的典型对抗性例子也可以用作评价模型强健性的工具。然而,在黑箱环境中,对抗性攻击的成功率可以进一步提高。因此,本研究将修改过的亚当梯度下行算法与迭代梯度攻击方法结合起来。随后,拟议的亚当超常快速渐进法被用于改善对抗性例子的可转移性。图象网的广泛实验表明,拟议的方法提供了比现有迭接方法更高的攻击性成功率。通过扩展我们的方法,我们在防御模型上取得了95.0%的先进攻击成功率。