Fooling deep neural networks (DNNs) with the black-box optimization has become a popular adversarial attack fashion, as the structural prior knowledge of DNNs is always unknown. Nevertheless, recent black-box adversarial attacks may struggle to balance their attack ability and visual quality of the generated adversarial examples (AEs) in tackling high-resolution images. In this paper, we propose an attention-guided black-box adversarial attack based on the large-scale multiobjective evolutionary optimization, termed as LMOA. By considering the spatial semantic information of images, we firstly take advantage of the attention map to determine the perturbed pixels. Instead of attacking the entire image, reducing the perturbed pixels with the attention mechanism can help to avoid the notorious curse of dimensionality and thereby improves the performance of attacking. Secondly, a large-scale multiobjective evolutionary algorithm is employed to traverse the reduced pixels in the salient region. Benefiting from its characteristics, the generated AEs have the potential to fool target DNNs while being imperceptible by the human vision. Extensive experimental results have verified the effectiveness of the proposed LMOA on the ImageNet dataset. More importantly, it is more competitive to generate high-resolution AEs with better visual quality compared with the existing black-box adversarial attacks.
翻译:以黑盒优化方式愚弄深层神经网络(DNNs)已经成为一种流行的对抗性攻击方式,因为DNNs的结构性先前知识总是未知的。然而,最近的黑盒对抗性攻击可能难以在应对高分辨率图像时平衡其攻击能力与生成的对抗性例子(AEs)的视觉质量。在本文中,我们建议以大规模多目标进化优化(称为LMOA)为基础,进行引人注意的黑盒对抗性攻击。考虑到图像的空间语义信息,产生的AE有可能愚弄DNes,同时无法被人类的黑洞察到。广泛的实验结果不是攻击整个图像,而是将接合的螺旋像素与关注机制一起减少,这样可以帮助避免臭名化的对立面图像的诅咒,从而改善攻击的性能。第二,我们用大规模多目标进化算法来绕过显要区域的减少的像素。从它的特性中受益,产生的AE有可能欺骗DNFs,而同时又无法被黑洞察到黑镜像像象。广泛的实验结果比起来更具有更高的竞争性性,A-BMOA 与较具有更高的视觉质量。