Neural network (NN) classifiers are vulnerable to adversarial attacks. Although the existing gradient-based attacks achieve state-of-the-art performance in feed-forward NNs and image recognition tasks, they do not perform as well on time series classification with recurrent neural network (RNN) models. This is because the cyclical structure of RNN prevents direct model differentiation and the visual sensitivity of time series data to perturbations challenges the traditional local optimization objective of the adversarial attack. In this paper, a black-box method called TSFool is proposed to efficiently craft highly-imperceptible adversarial time series for RNN classifiers. We propose a novel global optimization objective named Camouflage Coefficient to consider the imperceptibility of adversarial samples from the perspective of class distribution, and accordingly refine the adversarial attack as a multi-objective optimization problem to enhance the perturbation quality. To get rid of the dependence on gradient information, we also propose a new idea that introduces a representation model for RNN to capture deeply embedded vulnerable samples having otherness between their features and latent manifold, based on which the optimization solution can be heuristically approximated. Experiments on 10 UCR datasets are conducted to confirm that TSFool averagely outperforms existing methods with a 46.3% higher attack success rate, 87.4% smaller perturbation and 25.6% better Camouflage Coefficient at a similar time cost.
翻译:神经网络(NN)分类器容易受到对抗攻击。尽管现有的基于梯度的攻击在前馈NN和图像识别任务方面取得了最高的性能,但它们在循环神经网络(RNN)模型的时间序列分类方面表现不佳。这是因为RNN的循环结构防止直接模型分化,时间序列数据对扰动的视觉敏感性挑战了对抗攻击的传统局部优化目标。在本文中,我们提出了一种称为TSFool的黑盒方法,以高效制造用于RNN分类器的高度不可察觉的对抗时间序列。我们提出了一个新颖的全局优化目标,名为Camouflage Coefficient,它从类分布的角度考虑对抗样本的不可察觉性,在此基础上将对抗攻击作为多目标优化问题进行改进,从而增强扰动质量。为了摆脱对梯度信息的依赖,我们还提出了一种新的思路,引入了一个表示模型,可以捕捉具有特征和潜在流形之间的异质性的深度嵌入易受攻击的样本,从而可以启发式地近似优化解决方案。在10个UCR数据集上进行的实验证实,TSFool平均比现有方法具有46.3%的更高攻击成功率,87.4%的更小扰动和25.6%的更好Camouflage系数,时间成本相似。