Face recognition has achieved great success in the last five years due to the development of deep learning methods. However, deep convolutional neural networks (DCNNs) have been found to be vulnerable to adversarial examples. In particular, the existence of transferable adversarial examples can severely hinder the robustness of DCNNs since this type of attacks can be applied in a fully black-box manner without queries on the target system. In this work, we first investigate the characteristics of transferable adversarial attacks in face recognition by showing the superiority of feature-level methods over label-level methods. Then, to further improve transferability of feature-level adversarial examples, we propose DFANet, a dropout-based method used in convolutional layers, which can increase the diversity of surrogate models and obtain ensemble-like effects. Extensive experiments on state-of-the-art face models with various training databases, loss functions and network architectures show that the proposed method can significantly enhance the transferability of existing attack methods. Finally, by applying DFANet to the LFW database, we generate a new set of adversarial face pairs that can successfully attack four commercial APIs without any queries. This TALFW database is available to facilitate research on the robustness and defense of deep face recognition.
翻译:近五年来,由于深层次学习方法的发展,面对面的承认取得了巨大成功,然而,深层神经神经网络(DCNN)的深度演进很容易受到对抗性实例的影响,特别是,可转移的对抗性实例的存在会严重妨碍DCNN的稳健性,因为这种类型的攻击可以完全黑箱方式进行,而不必询问目标系统。在这项工作中,我们首先通过展示特征级方法优于标签级方法的可转移性来调查可转移的对抗性攻击的特征。然后,为了进一步提高地级对抗性实例的可转移性,我们建议DFANet,这是在同层中使用的一种基于辍学的方法,它可以增加代用模型的多样性,并获得同共性效果。在各种培训数据库、损失功能和网络结构中进行的关于状态式脸模型的广泛实验表明,拟议的方法可以大大提高现有攻击方法的可转移性。最后,通过将DFANet应用于LW数据库,我们制作了一套新的对抗性对面对面的对立式,可以成功地打击四套具有牢固的ALFA的防御性研究。