Deep learning is getting more and more outstanding performance in many tasks such as autonomous driving and face recognition and also has been challenged by different kinds of attacks. Adding perturbations that are imperceptible to human vision in an image can mislead the neural network model to get wrong results with high confidence. Adversarial Examples are images that have been added with specific noise to mislead a deep neural network model However, adding noise to images destroys the original data, making the examples useless in digital forensics and other fields. To prevent illegal or unauthorized access of image data such as human faces and ensure no affection to legal use reversible adversarial attack technique is rise. The original image can be recovered from its reversible adversarial example. However, the existing reversible adversarial examples generation strategies are all designed for the traditional imperceptible adversarial perturbation. How to get reversibility for locally visible adversarial perturbation? In this paper, we propose a new method for generating reversible adversarial examples based on local visual adversarial perturbation. The information needed for image recovery is embedded into the area beyond the adversarial patch by reversible data hiding technique. To reduce image distortion and improve visual quality, lossless compression and B-R-G embedding principle are adopted. Experiments on ImageNet dataset show that our method can restore the original images error-free while ensuring the attack performance.
翻译:深层学习在许多任务(如自主驱动和面部识别)中越来越表现出色,而且受到不同种类攻击的挑战。在图像中,对人的视觉无法察觉的额外扰动可能会误导神经网络模型,从而以高度自信获得错误的结果。反向实例是添加了特定噪音的图像,以误导深神经网络模型。在图像中添加噪音,破坏原始数据,使数字法证和其他领域的例子变得毫无用处。为了防止非法或未经授权获取图像数据(如人脸部),并确保对合法使用可逆对立攻击技术不感兴趣,正在上升。原始图像可以从可逆的对立攻击示例中恢复。然而,现有的可逆对立网络生成示例战略都是为传统的不可测的对立性对立干扰模型设计的。如何让本地可见的对立性对立性触动性触动性破坏原始数据的反弹性能?在本文中,我们提出了一种新的方法来生成可逆性对立性对立性对立性图像的对立性实例。图像复原所需的信息被嵌入到可逆性对立性对立原则以外的区域,而可变的对立性平的图像存储性图像则通过存储性变性图像显示性图像。