Deep hashing has shown promising performance in large-scale image retrieval. However, latent codes extracted by Deep Neural Networks (DNNs) will inevitably lose semantic information during the binarization process, which damages the retrieval accuracy and makes it challenging. Although many existing approaches perform regularization to alleviate quantization errors, we figure out an incompatible conflict between metric learning and quantization learning. The metric loss penalizes the inter-class distances to push different classes unconstrained far away. Worse still, it tends to map the latent code deviate from ideal binarization point and generate severe ambiguity in the binarization process. Based on the minimum distance of the binary linear code, we creatively propose Hashing-guided Hinge Function (HHF) to avoid such conflict. In detail, the carefully-designed inflection point, which relies on the hash bit length and category numbers, is explicitly adopted to balance the metric term and quantization term. Such a modification prevents the network from falling into local metric optimal minima in deep hashing. Extensive experiments in CIFAR-10, CIFAR-100, ImageNet, and MS-COCO show that HHF consistently outperforms existing techniques, and is robust and flexible to transplant into other methods. Code is available at https://github.com/JerryXu0129/HHF.
翻译:然而,深神经网络(DNNS)提取的潜在代码在二进制过程中将不可避免地失去语义信息,从而损害检索的准确性,并使其具有挑战性。虽然许多现有方法都实行正规化,以缓解量化错误,但我们发现,在计量学习和量化学习之间存在着不相容的冲突。衡量损失惩罚了将不同阶级推向不受到限制的距离。更糟糕的是,它往往映射潜代代码偏离理想的二进制点,并在二进制进程中产生严重的模糊性。基于二进制线代码的最低距离,我们创造性地提议Hashing-指导 Hinge 函数(HHHHF)以避免这种冲突。详细说来,我们仔细设计了透析点,该点依靠粗长的长度和类别数字,明确用于平衡计量术语和量化术语。这样的修改使得网络无法在深的集成中跌落到本地指标最佳迷你。 CIRF-10, CIFAR-100, IMFAR-29, image-HMFAR-GVER-GMVSMVS/ 和MS-COFORFORMSDFSMS/FSDSDFS/FODFORDS/FLVS/FOVDRVDRVDRVDRVDRVDS/M/MVDRVOVS/MGVOVOVOVOVOV 和NSOFOVF/MGFGFGFGFDFDFGVDFGVDFDFDFDFDFDFGFGFGVDR/MS/MGVDFDFDFGVDFD/MS/MSMSMSMSMSMDFGFGVDFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFGFG