Deep neural networks are vulnerable to adversarial examples that mislead models with imperceptible perturbations. In audio, although adversarial examples have achieved incredible attack success rates on white-box settings and black-box settings, most existing adversarial attacks are constrained by the input length. A More practical scenario is that the adversarial examples must be clipped or self-spliced and input into the black-box model. Therefore, it is necessary to explore how to improve transferability in different input length settings. In this paper, we take the synthetic speech detection task as an example and consider two representative SOTA models. We observe that the gradients of fragments with the same sample value are similar in different models via analyzing the gradients obtained by feeding samples into the model after cropping or self-splicing. Inspired by the above observation, we propose a new adversarial attack method termed sliding attack. Specifically, we make each sampling point aware of gradients at different locations, which can simulate the situation where adversarial examples are input to black-box models with varying input lengths. Therefore, instead of using the current gradient directly in each iteration of the gradient calculation, we go through the following three steps. First, we extract subsegments of different lengths using sliding windows. We then augment the subsegments with data from the adjacent domains. Finally, we feed the sub-segments into different models to obtain aggregate gradients to update adversarial examples. Empirical results demonstrate that our method could significantly improve the transferability of adversarial examples after clipping or self-splicing. Besides, our method could also enhance the transferability between models based on different features.
翻译:深心神经网络很容易受到敌对性例子的伤害,这些例子误导了不易察觉到的扰动模式。在音频方面,虽然对抗性例子在白箱设置和黑盒设置中取得了惊人的攻击成功率,但大多数现有的对抗性攻击受到输入长度的限制。更实际的假设是,敌对性例子必须剪裁或自冲并输入黑盒模型。因此,有必要探索如何在不同输入长度设置中改进可转移性。在本文中,我们以合成语音探测任务为例,考虑两个具有代表性的SOTA模型。我们观察到,在不同的模型中,具有相同抽样值的碎片梯度在白色框设置和黑盒设置中取得了惊人的攻击成功率,但通过分析将样品输入模型输入模型的梯度在裁剪辑或自我渗透后,我们提出了一个新的对抗性攻击方法,称为滑动攻击。具体地,我们让每个取样点了解不同地点的梯度,可以模拟在黑盒模型中输入不同输入的时间长度的模型的情况。因此,我们发现,不同样本中每个样本的梯度的梯度的梯度的梯度的梯度特点在不同的模型中是相似性,我们用不同的梯度更新结果的梯度计算,然后用不同的梯度计算,我们用不同的梯度计算,然后用不同的梯度转换的梯度转换的梯度转换的梯度转换到不同的梯度。我们用不同的梯度计算,我们用不同的梯度的梯度的梯度的梯度转换到不同的梯度计算,然后用不同的梯度的梯度计算,然后用不同的梯度转换到不同的梯度。