Transferability of adversarial examples is a key issue to apply this kind of attacks against multimedia forensics (MMF) techniques based on Deep Learning (DL) in a real-life setting. Adversarial example transferability, in fact, would open the way to the deployment of successful counter forensics attacks also in cases where the attacker does not have a full knowledge of the to-be-attacked system. Some preliminary works have shown that adversarial examples against CNN-based image forensics detectors are in general non-transferrable, at least when the basic versions of the attacks implemented in the most popular libraries are adopted. In this paper, we introduce a general strategy to increase the strength of the attacks and evaluate their transferability when such a strength varies. We experimentally show that, in this way, attack transferability can be largely increased, at the expense of a larger distortion. Our research confirms the security threats posed by the existence of adversarial examples even in multimedia forensics scenarios, thus calling for new defense strategies to improve the security of DL-based MMF techniques.
翻译:对抗性例子的可转让性是在现实生活中应用这种基于深层学习的多媒体法医学技术的这类攻击的一个关键问题。事实上,反性例子的可转让性将为成功部署反法医学攻击开辟道路,如果攻击者对即将攻击的系统并不完全了解,攻击者对将要攻击的系统也不十分了解。一些初步工作表明,针对有线电视新闻网图像法医学探测器的对抗性例子一般是不可转让的,至少当最受欢迎的图书馆实施的攻击的基本版本被采纳时是如此。在本文件中,我们引入了一种一般性战略,以提高攻击的强度,并在这种强度不同时评估其可转让性。我们实验性地表明,以这种方式,攻击性在很大程度上可以增加,而牺牲更大的扭曲性。我们的研究证实了即使在多媒体法医学假设中存在对抗性例子所造成的安全威胁,因此呼吁采取新的防御战略来改善以DL为基础的MF技术的安全。