The literature on adversarial attacks in computer vision typically focuses on pixel-level perturbations. These tend to be very difficult to interpret. Recent work that manipulates the latent representations of image generators to create "feature-level" adversarial perturbations gives us an opportunity to explore perceptible, interpretable adversarial attacks. We make three contributions. First, we observe that feature-level attacks provide useful classes of inputs for studying representations in models. Second, we show that these adversaries are uniquely versatile and highly robust. We demonstrate that they can be used to produce targeted, universal, disguised, physically-realizable, and black-box attacks at the ImageNet scale. Third, we show how these adversarial images can be used as a practical interpretability tool for identifying bugs in networks. We use these adversaries to make predictions about spurious associations between features and classes which we then test by designing "copy/paste" attacks in which one natural image is pasted into another to cause a targeted misclassification. Our results suggest that feature-level attacks are a promising approach for rigorous interpretability research. They support the design of tools to better understand what a model has learned and diagnose brittle feature associations. Code is available at https://github.com/thestephencasper/feature_level_adv
翻译:关于计算机视觉中的对抗性攻击的文献通常侧重于像素级的扰动。 这些情况往往很难解释。 最近利用图像生成器的潜在显示方式来制造“ 功能级” 对抗性扰动, 使我们有机会探索可以理解的、 可解释的对抗性攻击。 我们做出三点贡献。 首先, 我们观察到, 特征级攻击为研究模型中的演示提供了有用的投入类别。 第二, 我们显示, 这些对手具有独特的多才多才多艺, 并且非常强大。 我们表明, 这些对手可以用来在图像网络的尺度上产生有针对性的、 普遍、 变相的、 可实现的和黑盒攻击。 第三, 我们展示如何将这些对抗性图像用作识别网络中的错误的实用可解释工具。 我们利用这些对手来预测特征和类别之间的虚假联系,然后我们设计“ 复制/ 帕斯特” 攻击, 将其中一种自然图像粘贴到另一个模型, 导致有针对性的分类错误。 我们的结果表明, 地级攻击是一种很有希望的方法来进行严格的解释性的研究。 第三, 我们展示这些对抗性图像的图像可以用作一种实用的工具设计工具, 更好的理解/ 标准级 。