Recent studies have shown that neural network (NN) based image classifiers are highly vulnerable to adversarial examples, which poses a threat to security-sensitive image recognition task. Prior work has shown that JPEG compression can combat the drop in classification accuracy on adversarial examples to some extent. But, as the compression ratio increases, traditional JPEG compression is insufficient to defend those attacks but can cause an abrupt accuracy decline to the benign images. In this paper, with the aim of fully filtering the adversarial perturbations, we firstly make modifications to traditional JPEG compression algorithm which becomes more favorable for NN. Specifically, based on an analysis of the frequency coefficient, we design a NN-favored quantization table for compression. Considering compression as a data augmentation strategy, we then combine our model-agnostic preprocess with noisy training. We fine-tune the pre-trained model by training with images encoded at different compression levels, thus generating multiple classifiers. Finally, since lower (higher) compression ratio can remove both perturbations and original features slightly (aggressively), we use these trained multiple models for model ensemble. The majority vote of the ensemble of models is adopted as final predictions. Experiments results show our method can improve defense efficiency while maintaining original accuracy.
翻译:最近的研究显示,基于神经网络(NN)的图像分类者极易受到对抗性实例的伤害,这对安全敏感图像识别任务构成威胁。先前的工作显示,JPEG压缩可以在某种程度上克服对抗性实例分类准确性下降的问题。但是,随着压缩率的提高,传统的JPEG压缩方法不足以防御这些攻击,但可能导致良性图像的突然精确性下降。在本文件中,为了充分过滤对立性扰动,我们首先修改传统的JPEG压缩算法,这种算法对NNN更有利。具体地说,根据对频率系数的分析,我们设计了一个NPEG压缩表,将NEG压缩方法作为数据增强战略考虑,然后将我们的模型-不可知性前期程序与噪音培训结合起来。我们通过在不同压缩级别编码的图像对培训前模型进行微调,从而产生多重分类器。最后(高)压缩比率可以略地消除对NNEN的过错和原始特征。我们使用这些经过训练的多模型进行压缩的压缩。我们使用这些经过训练的模型来进行压缩,作为数据增强战略的精确性预测。我们原始的实验性的方法可以显示原始的精确性方法。