Aliasing is a highly important concept in signal processing, as careful consideration of resolution changes is essential in ensuring transmission and processing quality of audio, image, and video. Despite this, up until recently aliasing has received very little consideration in Deep Learning, with all common architectures carelessly sub-sampling without considering aliasing effects. In this work, we investigate the hypothesis that the existence of adversarial perturbations is due in part to aliasing in neural networks. Our ultimate goal is to increase robustness against adversarial attacks using explainable, non-trained, structural changes only, derived from aliasing first principles. Our contributions are the following. First, we establish a sufficient condition for no aliasing for general image transformations. Next, we study sources of aliasing in common neural network layers, and derive simple modifications from first principles to eliminate or reduce it. Lastly, our experimental results show a solid link between anti-aliasing and adversarial attacks. Simply reducing aliasing already results in more robust classifiers, and combining anti-aliasing with robust training out-performs solo robust training on $L_2$ attacks with none or minimal losses in performance on $L_{\infty}$ attacks.
翻译:在信号处理中,一个非常重要的概念是异质处理,因为仔细考虑分辨率变化对于确保音频、图像和视频的传输和处理质量至关重要。尽管如此,直到最近,在深地学习中,直到最近之前的别名很少得到什么考虑,所有共同结构都粗略地进行子抽样,而没有考虑别名的效果。在这项工作中,我们调查了这样一种假设,即对抗性扰动的存在部分是由于神经网络中的化名造成的。我们的最终目标是,利用更强有力的分级器和从别名中推导出的唯一结构变化,提高对抗性攻击的稳健性。我们的贡献如下:首先,我们为一般图像转换不作别名创造了充分的条件。接下来,我们研究共同神经网络层中的别名,并从最初的原则中简单修改消除或减少它。最后,我们的实验结果显示,反报复性攻击与对抗性攻击之间有着牢固的联系。简单地减少已经通过更强有力的分级化的、不经过训练的结果,并且将反欺诈与强力的以美元为单位的反性攻击损失结合起来。