Face recognition is greatly improved by deep convolutional neural networks (CNNs). Recently, these face recognition models have been used for identity authentication in security sensitive applications. However, deep CNNs are vulnerable to adversarial patches, which are physically realizable and stealthy, raising new security concerns on the real-world applications of these models. In this paper, we evaluate the robustness of face recognition models using adversarial patches based on transferability, where the attacker has limited accessibility to the target models. First, we extend the existing transfer-based attack techniques to generate transferable adversarial patches. However, we observe that the transferability is sensitive to initialization and degrades when the perturbation magnitude is large, indicating the overfitting to the substitute models. Second, we propose to regularize the adversarial patches on the low dimensional data manifold. The manifold is represented by generative models pre-trained on legitimate human face images. Using face-like features as adversarial perturbations through optimization on the manifold, we show that the gaps between the responses of substitute models and the target models dramatically decrease, exhibiting a better transferability. Extensive digital world experiments are conducted to demonstrate the superiority of the proposed method in the black-box setting. We apply the proposed method in the physical world as well.
翻译:深层神经神经网络(CNNs)大大改善了对面面的识别。最近,这些面部识别模型在安全敏感应用中被用于身份认证。然而,深重CNN很容易受到对抗性补丁的伤害,这些补丁在物理上可实现且隐形,这些补丁在实际应用这些模型的现实世界中提出了新的安全关切。在本文中,我们评估了以可转移性为基础的对抗性补丁的面部识别模型的稳健性,攻击者在可转移性基础上对目标模型的可获取性进行了限制。首先,我们扩大了现有的基于转让的攻击技术,以产生可转移的对面补丁。然而,我们观察到,当扰动性强度大时,这种可转移性对初始化和降解十分敏感,这表明对替代模型的过度适用。第二,我们建议对低维数据多重的对抗性补丁进行规范。我们用对合法的人类脸部图像进行预先训练的变形模型代表了这些方形模型。通过对方位进行优化,将面部位的特征作为对抗性穿孔。我们发现替代模型与目标模型之间存在的差距大大缩小,我们展示了替代模型的反应,显示替代模型与目标模型之间的差距,展示了黑的可转移性。我们所拟采用的方法。我们所展开的物理上的世界。