Deep Neural Networks (DNNs) are highly sensitive to imperceptible malicious perturbations, known as adversarial attacks. Following the discovery of this vulnerability in real-world imaging and vision applications, the associated safety concerns have attracted vast research attention, and many defense techniques have been developed. Most of these defense methods rely on adversarial training (AT) -- training the classification network on images perturbed according to a specific threat model, which defines the magnitude of the allowed modification. Although AT leads to promising results, training on a specific threat model fails to generalize to other types of perturbations. A different approach utilizes a preprocessing step to remove the adversarial perturbation from the attacked image. In this work, we follow the latter path and aim to develop a technique that leads to robust classifiers across various realizations of threat models. To this end, we harness the recent advances in stochastic generative modeling, and means to leverage these for sampling from conditional distributions. Our defense relies on an addition of Gaussian i.i.d noise to the attacked image, followed by a pretrained diffusion process -- an architecture that performs a stochastic iterative process over a denoising network, yielding a high perceptual quality denoised outcome. The obtained robustness with this stochastic preprocessing step is validated through extensive experiments on the CIFAR-10 dataset, showing that our method outperforms the leading defense methods under various threat models.
翻译:深心神经网络(DNNS)对于不可察觉的恶意扰动非常敏感。 在现实世界成像和视觉应用中发现这种脆弱性后,相关的安全关切引起了广泛的研究关注,并开发了许多防御技术。这些防御方法大多依赖于对抗性培训(AT) -- -- 根据特定的威胁模型对图像进行被扰动的分类网络培训,该模型界定了允许修改的程度。虽然AT导致有希望的结果,但关于特定威胁模型的培训未能推广到其他类型的扰动。不同的方法利用预处理步骤消除被攻击图像中的对抗性扰动。在这项工作中,我们遵循后一种途径,目的是开发一种技术,使各种认识威胁模型的各方能够进行稳健的分类。为此,我们利用了在随机分析分析模型中的最新进展,从有条件分布中将这些进展用于取样。我们的防御依赖于在被攻击图像中添加高估的一.i. 噪音,然后是前期扩散过程。我们遵循的是后期的高度扩散过程。我们遵循后期的后期防御系统结构将展示一种稳健的系统结果。