Learning compact binary codes for image retrieval problem using deep neural networks has attracted increasing attention recently. However, training deep hashing networks is challenging due to the binary constraints on the hash codes, the similarity preserving property, and the requirement for a vast amount of labelled images. To the best of our knowledge, none of the existing methods has tackled all of these challenges completely in a unified framework. In this work, we propose a novel end-to-end deep hashing approach, which is trained to produce binary codes directly from image pixels without the need of manual annotation. In particular, we propose a novel pairwise binary constrained loss function, which simultaneously encodes the distances between pairs of hash codes, and the binary quantization error. In order to train the network with the proposed loss function, we also propose an efficient parameter learning algorithm. In addition, to provide similar/dissimilar training images to train the network, we exploit 3D models reconstructed from unlabelled images for automatic generation of enormous similar/dissimilar pairs. Extensive experiments on three image retrieval benchmark datasets demonstrate the superior performance of the proposed method over the state-of-the-art hashing methods on the image retrieval problem.
翻译:使用深神经网络的图像检索问题学习压缩二进制代码最近引起越来越多的关注。然而,由于散列代码、相似保护属性和大量贴标签图像要求的二进制限制,深散列网络培训具有挑战性。据我们所知,现有方法中没有任何一种方法在一个统一的框架内完全解决所有这些挑战。在这项工作中,我们建议采用一种新的端对端深海散列法,这种方法经过培训,直接从图像像素中生成二进制代码,无需人工批注。特别是,我们提议了一个新颖的双对式二进制限制损失功能,同时将散列码和二进制分解错误之间的距离编码。为了用拟议的损失函数来培训网络,我们还提议了一个高效的参数学习算法。此外,为了提供类似的/不同的培训图像来培训网络,我们利用从未贴标签的图像中重建的3D模型来自动生成巨大的相似/不同配对。我们在三个图像基准数据集上进行广泛的实验,这三张图像检索方法展示了在状态上的拟议图像检索方法的优劣性。