Machine learning models are critically susceptible to evasion attacks from adversarial examples. Generally, adversarial examples, modified inputs deceptively similar to the original input, are constructed under whitebox settings by adversaries with full access to the model. However, recent attacks have shown a remarkable reduction in query numbers to craft adversarial examples using blackbox attacks. Particularly, alarming is the ability to exploit the classification decision from the access interface of a trained model provided by a growing number of Machine Learning as a Service providers including Google, Microsoft, IBM and used by a plethora of applications incorporating these models. The ability of an adversary to exploit only the predicted label from a model to craft adversarial examples is distinguished as a decision-based attack. In our study, we first deep dive into recent state-of-the-art decision-based attacks in ICLR and SP to highlight the costly nature of discovering low distortion adversarial employing gradient estimation methods. We develop a robust query efficient attack capable of avoiding entrapment in a local minimum and misdirection from noisy gradients seen in gradient estimation methods. The attack method we propose, RamBoAttack, exploits the notion of Randomized Block Coordinate Descent to explore the hidden classifier manifold, targeting perturbations to manipulate only localized input features to address the issues of gradient estimation methods. Importantly, the RamBoAttack is more robust to the different sample inputs available to an adversary and the targeted class. Overall, for a given target class, RamBoAttack is demonstrated to be more robust at achieving a lower distortion within a given query budget. We curate our extensive results using the large-scale high-resolution ImageNet dataset and open-source our attack, test samples and artifacts on GitHub.
翻译:机器学习模式极易受到来自对抗性实例的攻击。 一般来说, 对抗性实例、 与原始输入相似的修改性投入, 是由对手在白箱设置下建造的, 并且完全可以使用该模型。 然而, 最近的攻击显示, 使用黑箱攻击, 利用黑箱攻击 来编造对抗性攻击 。 特别令人震惊的是, 能够利用由越来越多的机器学习机构作为服务供应商提供的经过训练的模型的准入界面作出的分类决定, 包括谷歌、 微软、 IBM 和大量应用包含这些模型的应用。 对手仅利用模型的预测标签标签来制作对抗性例子的能力被区别为基于决定的攻击。 在我们的研究中, 我们首先深入潜入最近最先进的基于黑箱攻击性攻击性攻击性的例子, 来突出使用梯度估计方法发现低扭曲性对抗性对抗性攻击性攻击性攻击性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定性决定