Deep hashing has been extensively utilized in massive image retrieval because of its efficiency and effectiveness. However, deep hashing models are vulnerable to adversarial examples, making it essential to develop adversarial defense methods for image retrieval. Existing solutions achieved limited defense performance because of using weak adversarial samples for training and lacking discriminative optimization objectives to learn robust features. In this paper, we present a min-max based Center-guided Adversarial Training, namely CgAT, to improve the robustness of deep hashing networks through worst adversarial examples. Specifically, we first formulate the center code as a semantically-discriminative representative of the input image content, which preserves the semantic similarity with positive samples and dissimilarity with negative examples. We prove that a mathematical formula can calculate the center code immediately. After obtaining the center codes in each optimization iteration of the deep hashing network, they are adopted to guide the adversarial training process. On the one hand, CgAT generates the worst adversarial examples as augmented data by maximizing the Hamming distance between the hash codes of the adversarial examples and the center codes. On the other hand, CgAT learns to mitigate the effects of adversarial samples by minimizing the Hamming distance to the center codes. Extensive experiments on the benchmark datasets demonstrate the effectiveness of our adversarial training algorithm in defending against adversarial attacks for deep hashing-based retrieval. Compared with the current state-of-the-art defense method, we significantly improve the defense performance by an average of 18.61%, 12.35%, and 11.56% on FLICKR-25K, NUS-WIDE, and MS-COCO, respectively.
翻译:在大规模图像检索中广泛使用了深 hashing 深深层 hashing, 因为它的效率和效力。 但是, 深层 hash 模型很容易成为对抗性例子, 使得开发对抗性防御方法对图像检索至关重要。 现有的解决方案由于在培训中使用了较弱的对抗性样本, 并且缺乏区别性优化目标来学习强性特征, 因而取得了有限的防御性表现。 在本文中, 我们展示了一个基于 min-max 的基于中心指导的反向培训, 即CgAT, 通过最差的对称性对称性培训实例, 来提高深度的对称网络的稳健性。 具体地说, 我们首先将中心代码作为输入图像内容内容的语义差异性代表, 使语义相似性与正样相似, 与负面例子不同。 我们证明数学公式可以立即计算出中心代码。 在每次优化深处的网络中, 获得中心代码后, 就会被采用。 一方面, CgAT 生成了最差的对立性美国对称的例子, 通过尽可能扩大的数据,, 我们通过 Hammrefrefrial- detradefrefor devideferal 的防御 Cradeformation, 在11 的 Chardeformal- devald 的代码中, 的 Charding defreme ex ex ex ex prevation ex ex exbal ex ex exprevation ex express exbal exprevation,, pressal ex ex ex ex ex exx, lex leg sal leg sal lear- sal- sal- sal- sal- sal- sal- sal- sal- sal- sal- sal- sal- sal- sal- legal- sal- sal- sal- sal- sal- sal- sal- legal- sal- sal- sal- sal- sal- sal- sal- sal- lear- sal- sal- sal- sal- sal- sal-