Deep neural networks are vulnerable to adversarial examples, which are crafted by applying small, human-imperceptible perturbations on the original images, so as to mislead deep neural networks to output inaccurate predictions. Adversarial attacks can thus be an important method to evaluate and select robust models in safety-critical applications. However, under the challenging black-box setting, most existing adversarial attacks often achieve relatively low success rates on adversarially trained networks and advanced defense models. In this paper, we propose AdaBelief Iterative Fast Gradient Method (ABI-FGM) and Crop-Invariant attack Method (CIM) to improves the transferability of adversarial examples. ABI-FGM and CIM can be readily integrated to build a strong gradient-based attack to further boost the success rates of adversarial examples for black-box attacks. Moreover, our method can also be naturally combined with other gradient-based attack methods to build a more robust attack to generate more transferable adversarial examples against the defense models. Extensive experiments on the ImageNet dataset demonstrate the method's effectiveness. Whether on adversarially trained networks or advanced defense models, our method has higher success rates than state-of-the-art gradient-based attack methods.
翻译:深心神经网络容易受到对抗性例子的伤害,这些例子是通过在原始图像上应用小的、人类无法察觉的干扰而形成的,目的是误导深心神经网络,以产生不准确的预测。因此,反向攻击可能是评估和选择安全关键应用中强健模型的一个重要方法。然而,在挑战性的黑箱设置下,大多数现有的对抗性攻击往往在对抗性训练的网络和先进的防御模型上达到相对较低的成功率。在本文中,我们提议Adabelief 热性快速渐进法(ABI-FGM)和作物性差异性攻击方法(CIM)改进对抗性例子的可转移性。ABI-FGM和CIM可以很容易地整合成一个强大的梯度攻击,以进一步提升黑箱攻击的对抗性例子的成功率。此外,我们的方法也可以自然地与其他基于梯度的攻击方法相结合,建立更强大的攻击性攻击性攻击,以产生更可转让的对抗性快速渐进式对抗性攻击模型。在图像网络上进行的广泛实验,展示了方法的有效性。无论是在对抗性攻击性攻击性攻击性攻击率的网络上,还是先进的防御性先进防御性攻击性攻击率模型上,都具有较高的防御性攻击性试验。