Despite their unmatched performance, deep neural networks remain susceptible to targeted attacks by nearly imperceptible levels of adversarial noise. While the underlying cause of this sensitivity is not well understood, theoretical analyses can be simplified by reframing each layer of a feed-forward network as an approximate solution to a sparse coding problem. Iterative solutions using basis pursuit are theoretically more stable and have improved adversarial robustness. However, cascading layer-wise pursuit implementations suffer from error accumulation in deeper networks. In contrast, our new method of deep pursuit approximates the activations of all layers as a single global optimization problem, allowing us to consider deeper, real-world architectures with skip connections such as residual networks. Experimentally, our approach demonstrates improved robustness to adversarial noise.
翻译:深层神经网络尽管表现不相称,但几乎无法察觉的对抗性噪音水平仍然容易成为有针对性的攻击目标。虽然这一敏感度的根本原因没有得到很好理解,但理论分析可以简化,办法是重新组合一个支线前网络的每一层,作为稀疏编码问题的大致解决办法。在理论上,利用基础追求的迭生解决办法比较稳定,并改进了对抗性强力。然而,分层追求的分层追求的实施在更深的网络中受到错误积累的影响。相比之下,我们新的深度追求方法将所有层次的启动近似为单一的全球优化问题,从而使我们能够考虑具有诸如残余网络等跳过连接的更深层、真实世界结构。实验性地说,我们的方法显示了对对抗性噪音的强大程度。