In hash-based image retrieval systems, degraded or transformed inputs usually generate different codes from the original, deteriorating the retrieval accuracy. To mitigate this issue, data augmentation can be applied during training. However, even if augmented samples of an image are similar in real feature space, the quantization can scatter them far away in Hamming space. This results in representation discrepancies that can impede training and degrade performance. In this work, we propose a novel self-distilled hashing scheme to minimize the discrepancy while exploiting the potential of augmented data. By transferring the hash knowledge of the weakly-transformed samples to the strong ones, we make the hash code insensitive to various transformations. We also introduce hash proxy-based similarity learning and binary cross entropy-based quantization loss to provide fine quality hash codes. Ultimately, we construct a deep hashing framework that not only improves the existing deep hashing approaches, but also achieves the state-of-the-art retrieval results. Extensive experiments are conducted and confirm the effectiveness of our work.
翻译:在基于散列的图像检索系统中,退化或转变输入通常产生与原始的、恶化的检索准确性不同的代码。为了缓解这一问题,可以在培训期间应用数据增强。但是,即使图像的放大样本在真实地物空间中类似,量化也可以在哈明州很远的地方散布。这会造成代表差异,从而可能妨碍培训和降低性能。在这项工作中,我们提出了一个新的自我蒸馏法,以在利用扩大的数据潜力的同时尽量减少差异。通过将低变异样品的散装知识转移给强者,我们使散装代码对各种变异不敏感。我们还引入了基于散列的代理类似性学习和二进式跨反倍的四倍化损失,以提供精细质质质的散变码。我们最终建立一个深度的散射框架,不仅改进现有的深散射法,而且实现最新检索结果。我们进行了广泛的实验,并证实了我们的工作的有效性。