Recent studies have highlighted that deep neural networks (DNNs) are vulnerable to adversarial attacks, even in a black-box scenario. However, most of the existing black-box attack algorithms need to make a huge amount of queries to perform attacks, which is not practical in the real world. We note one of the main reasons for the massive queries is that the adversarial example is required to be visually similar to the original image, but in many cases, how adversarial examples look like does not matter much. It inspires us to introduce a new attack called \emph{input-free} attack, under which an adversary can choose an arbitrary image to start with and is allowed to add perceptible perturbations on it. Following this approach, we propose two techniques to significantly reduce the query complexity. First, we initialize an adversarial example with a gray color image on which every pixel has roughly the same importance for the target model. Then we shrink the dimension of the attack space by perturbing a small region and tiling it to cover the input image. To make our algorithm more effective, we stabilize a projected gradient ascent algorithm with momentum, and also propose a heuristic approach for region size selection. Through extensive experiments, we show that with only 1,701 queries on average, we can perturb a gray image to any target class of ImageNet with a 100\% success rate on InceptionV3. Besides, our algorithm has successfully defeated two real-world systems, the Clarifai food detection API and the Baidu Animal Identification API.
翻译:最近的研究表明,深层神经网络(DNNS)即使在黑盒情景下也很容易受到对抗性攻击。然而,大多数现有的黑盒攻击算法需要大量查询才能进行攻击,这在现实世界中是不切实际的。我们注意到,大规模查询的主要原因之一是,敌对性实例需要与原始图像相近,但在许多情况下,对抗性实例看起来并不重要。它激励我们引入了名为\emph{input-fret}攻击的新攻击,在这种攻击下,一个对手可以选择一个可以启动的任意图像,并允许增加可以察觉到的扰动。按照这种方法,我们建议了两种技术来大幅降低查询的复杂性。首先,我们开始了一个带有灰色图像的对抗性例子,每个像素的图像对目标模型都具有同样的重要性。然后,我们缩小了攻击空间的维度,通过一个小区域来破坏一个小区域,用钢筋来覆盖输入图像。为了提高我们的算法的有效性,我们要将一个预测的梯度稳定成灰色的图像,然后用一个平均速度来显示一个区域。