Despite our best efforts, deep learning models remain highly vulnerable to even tiny adversarial perturbations applied to the inputs. The ability to extract information from solely the output of a machine learning model to craft adversarial perturbations to black-box models is a practical threat against real-world systems, such as autonomous cars or machine learning models exposed as a service (MLaaS). Of particular interest are sparse attacks. The realization of sparse attacks in black-box models demonstrates that machine learning models are more vulnerable than we believe. Because these attacks aim to minimize the number of perturbed pixels measured by l_0 norm-required to mislead a model by solely observing the decision (the predicted label) returned to a model query; the so-called decision-based attack setting. But, such an attack leads to an NP-hard optimization problem. We develop an evolution-based algorithm-SparseEvo-for the problem and evaluate against both convolutional deep neural networks and vision transformers. Notably, vision transformers are yet to be investigated under a decision-based attack setting. SparseEvo requires significantly fewer model queries than the state-of-the-art sparse attack Pointwise for both untargeted and targeted attacks. The attack algorithm, although conceptually simple, is also competitive with only a limited query budget against the state-of-the-art gradient-based whitebox attacks in standard computer vision tasks such as ImageNet. Importantly, the query efficient SparseEvo, along with decision-based attacks, in general, raise new questions regarding the safety of deployed systems and poses new directions to study and understand the robustness of machine learning models.
翻译:尽管我们尽了最大努力,深层次的学习模式仍然极易受到对投入的微小对抗性扰动的影响。仅仅从一个机器学习模式的输出中提取信息,将对抗性扰动生成黑盒模式的能力,是对现实世界系统的实际威胁,例如自动汽车或机器学习模式,作为服务(MlaaAS)暴露。特别令人感兴趣的是少发攻击。在黑盒模型中发现少发的攻击表明机器学习模式比我们相信的要脆弱得多。由于这些攻击的目的是通过仅仅观察决定(预测标签)返回到模式查询,从而误导模型;所谓的基于决定的攻击设置。但是,这种攻击导致一个基于NP的硬性汽车或机器学习模式的优化问题。我们针对问题开发了一个基于进化的算法-SparseEvo,并且对革命性的深层神经网络网络和视觉变异器进行评估。值得注意的是,基于标准的视觉变异器尚未在基于决定的攻击设置中被调查。 精确的Eprevoral Evo要求模型的查询要大大少于以观察模型来测量攻击的模型,尽管有目标性攻击的精确性攻击, 也只是精确的精确的计算。