Capsule network is a type of neural network that uses the spatial relationship between features to classify images. By capturing the poses and relative positions between features, its ability to recognize affine transformation is improved, and it surpasses traditional convolutional neural networks (CNNs) when handling translation, rotation and scaling. The Stacked Capsule Autoencoder (SCAE) is the state-of-the-art capsule network. The SCAE encodes an image as capsules, each of which contains poses of features and their correlations. The encoded contents are then input into the downstream classifier to predict the categories of the images. Existing research mainly focuses on the security of capsule networks with dynamic routing or EM routing, and little attention has been given to the security and robustness of the SCAE. In this paper, we propose an evasion attack against the SCAE. After a perturbation is generated based on the output of the object capsules in the model, it is added to an image to reduce the contribution of the object capsules related to the original category of the image so that the perturbed image will be misclassified. We evaluate the attack using an image classification experiment, and the experimental results indicate that the attack can achieve high success rates and stealthiness. It confirms that the SCAE has a security vulnerability whereby it is possible to craft adversarial samples without changing the original structure of the image to fool the classifiers. We hope that our work will make the community aware of the threat of this attack and raise the attention given to the SCAE's security.
翻译:Capsule 网络是一种神经网络, 使用各功能之间的空间关系来分类图像。 通过捕捉各功能之间的配置和相对位置, 其辨别松动变异的能力得到提高, 并超越了处理翻译、 旋转和缩放的传统神经网络(CNNs CNNs ) 。 Stacked Capsule Autoencoder (SCAE) 是最先进的胶囊网络 。 SCAE 将图像编码成胶囊, 其中每个包含特性和相关性。 编码的内容随后被输入下游分类器, 以预测图像的类别。 现有的研究主要侧重于带有动态变换或EM 变换的胶囊网络的安全性, 并且很少关注 SCAE 的安全和稳健性 。 在根据模型中对象胶囊的输出产生扰动后, SCAE 将添加到一个图像中, 用于减少与原始图像类别相关的对象胶囊的贡献 。 我们对原始变变变变变变的图像 。