Image hashing is a principled approximate nearest neighbor approach to find similar items to a query in a large collection of images. Hashing aims to learn a binary-output function that maps an image to a binary vector. For optimal retrieval performance, producing balanced hash codes with low-quantization error to bridge the gap between the learning stage's continuous relaxation and the inference stage's discrete quantization is important. However, in the existing deep supervised hashing methods, coding balance and low-quantization error are difficult to achieve and involve several losses. We argue that this is because the existing quantization approaches in these methods are heuristically constructed and not effective to achieve these objectives. This paper considers an alternative approach to learning the quantization constraints. The task of learning balanced codes with low quantization error is re-formulated as matching the learned distribution of the continuous codes to a pre-defined discrete, uniform distribution. This is equivalent to minimizing the distance between two distributions. We then propose a computationally efficient distributional distance by leveraging the discrete property of the hash functions. This distributional distance is a valid distance and enjoys lower time and sample complexities. The proposed single-loss quantization objective can be integrated into any existing supervised hashing method to improve code balance and quantization error. Experiments confirm that the proposed approach substantially improves the performance of several representative hashing~methods.
翻译:图像 hash 是一个有原则的近邻近邻方法, 以找到与大量图像收藏中查询的类似项目。 散列的目的是学习一个二进制输出函数, 将图像映射成二进制矢量。 为了优化检索性能, 生成平衡的散列代码, 低量化错误, 以弥补学习阶段持续放松和推断阶段离散四分化之间的差距。 但是, 在现有的深层监督散列方法中, 编码平衡和低量化错误很难实现, 并包含一些损失。 我们争辩说, 这是因为这些方法中的现有量化方法是超自然构造的, 无法有效实现这些目标。 为了优化检索性能, 生成平衡的散列代码, 以低量化错误来弥补学习的平衡代码。 将连续代码的分布分布分布分布分布分配到预定义的离散分解、 统一分布。 我们然后提议通过利用 hash 函数的离散属性来计算高效的分布距离。 这个分布距离和缩略度方法是有效的, 运行一个有效的单一距离和最小化方法, 改进现有。