Deep neural networks (DNNs) have been shown to be vulnerable to adversarial examples. Moreover, the transferability of the adversarial examples has received broad attention in recent years, which means that adversarial examples crafted by a surrogate model can also attack unknown models. This phenomenon gave birth to the transfer-based adversarial attacks, which aim to improve the transferability of the generated adversarial examples. In this paper, we propose to improve the transferability of adversarial examples in the transfer-based attack via masking unimportant parameters (MUP). The key idea in MUP is to refine the pretrained surrogate models to boost the transfer-based attack. Based on this idea, a Taylor expansion-based metric is used to evaluate the parameter importance score and the unimportant parameters are masked during the generation of adversarial examples. This process is simple, yet can be naturally combined with various existing gradient-based optimizers for generating adversarial examples, thus further improving the transferability of the generated adversarial examples. Extensive experiments are conducted to validate the effectiveness of the proposed MUP-based methods.
翻译:深度神经网络(DNNs)已经被证明对对抗性样本具有弱点。此外,对抗样本的可转移性在最近几年受到了广泛的关注,这意味着由代理模型生成的对抗样本也可能攻击未知模型。这种现象产生了基于转移的对抗攻击,旨在提高生成的对抗样本的可转移性。在本文中,我们提出通过遮蔽不重要的参数(MUP)来改进转移攻击中对抗样本的可转移性。MUP中的关键思想是通过优化预训练的代理模型来提高转移攻击的效果。基于这个想法,使用基于泰勒展开的度量来评估参数重要性得分,并在生成对抗样本时遮蔽不重要的参数。这个过程是简单的,但可以自然地与各种现有的基于梯度的优化器相结合,从而进一步提高生成的对抗样本的可转移性。进行了大量实验来验证所提出的基于MUP的方法的有效性。