Deep Neural Network classifiers are vulnerable to adversarial attack, where an imperceptible perturbation could result in misclassification. However, the vulnerability of DNN-based image ranking systems remains under-explored. In this paper, we propose two attacks against deep ranking systems, i.e., Candidate Attack and Query Attack, that can raise or lower the rank of chosen candidates by adversarial perturbations. Specifically, the expected ranking order is first represented as a set of inequalities, and then a triplet-like objective function is designed to obtain the optimal perturbation. Conversely, an anti-collapse triplet defense is proposed to improve the ranking model robustness against all proposed attacks, where the model learns to prevent the positive and negative samples being pulled close to each other by adversarial attack. To comprehensively measure the empirical adversarial robustness of a ranking model with our defense, we propose an empirical robustness score, which involves a set of representative attacks against ranking models. Our adversarial ranking attacks and defenses are evaluated on MNIST, Fashion-MNIST, CUB200-2011, CARS196 and Stanford Online Products datasets. Experimental results demonstrate that a typical deep ranking system can be effectively compromised by our attacks. Nevertheless, our defense can significantly improve the ranking system robustness, and simultaneously mitigate a wide range of attacks.
翻译:深神经网分类系统很容易受到对抗性攻击, 由此可见的触动可能会导致错误分类。 但是, DNN 的图像排名系统仍然处于弱势状态。 在本文中,我们建议对深层次排名系统进行两次攻击, 即候选人攻击和询问攻击, 这些系统可以通过对抗性扰动提高或降低所选候选人的级别。 具体地说, 预期的排名顺序首先代表一系列不平等,然后设计三重相似的目标功能, 以获得最佳的扰动。 相反, 提议进行三重防御, 以提高所有拟议袭击的排名模型的稳健性, 在这些系统中,模型学会通过对抗性攻击来防止正面和负面的样本被相互拉近。 为了全面衡量与我们防御性对立的排名模型的对抗性稳健健性,我们提出了实证性评分, 包括对排名模型进行一系列有代表性的攻击。 我们的对抗性攻击和防御性攻击被评估为MNIST、 Fashon- MNIST、 CUB200- 2011、 CAR- 196 和斯坦标准级系统 能够大幅地改进我们的典型防御系统。