Adversarial attacks in deep learning models, especially for safety-critical systems, are gaining more and more attention in recent years, due to the lack of trust in the security and robustness of AI models. Yet the more primitive adversarial attacks might be physically infeasible or require some resources that are hard to access like the training data, which motivated the emergence of patch attacks. In this survey, we provide a comprehensive overview to cover existing techniques of adversarial patch attacks, aiming to help interested researchers quickly catch up with the progress in this field. We also discuss existing techniques for developing detection and defences against adversarial patches, aiming to help the community better understand this field and its applications in the real world.
翻译:近些年来,由于对AI模型的安全和稳健性缺乏信任,深层学习模式,特别是安全临界系统方面的对立攻击越来越受到越来越多的关注,然而,更原始的对抗性攻击在物理上可能不可行,或需要一些难以获取的资源,如导致补丁攻击的训练数据。在这次调查中,我们提供了全面的概览,以涵盖对抗性补丁攻击的现有技术,目的是帮助感兴趣的研究人员迅速赶上这一领域的进展。我们还讨论了针对对抗性补丁发展探测和防御的现有技术,目的是帮助社区更好地了解这一领域及其在现实世界中的应用。