Recently, deep hashing with Hamming distance metric has drawn increasing attention for face image retrieval tasks. However, its counterpart deep quantization methods, which learn binary code representations with dictionary-related distance metrics, have seldom been explored for the task. This paper makes the first attempt to integrate product quantization into an end-to-end deep learning framework for face image retrieval. Unlike prior deep quantization methods where the codewords for quantization are learned from data, we propose a novel scheme using predefined orthonormal vectors as codewords, which aims to enhance the quantization informativeness and reduce the codewords' redundancy. To make the most of the discriminative information, we design a tailored loss function that maximizes the identity discriminability in each quantization subspace for both the quantized and the original features. Furthermore, an entropy-based regularization term is imposed to reduce the quantization error. We conduct experiments on three commonly-used datasets under the settings of both single-domain and cross-domain retrieval. It shows that the proposed method outperforms all the compared deep hashing/quantization methods under both settings with significant superiority. The proposed codewords scheme consistently improves both regular model performance and model generalization ability, verifying the importance of codewords' distribution for the quantization quality. Besides, our model's better generalization ability than deep hashing models indicates that it is more suitable for scalable face image retrieval tasks.
翻译:最近, 与 Hamming 远程测量的深度散射让人们日益关注面部图像检索任务。 然而, 其对应的深度量化方法, 即学习与字典相关的远距测量的二进制代号表达式, 却很少为任务探索。 本文首次尝试将产品量化纳入一个端到端深学习框架, 供面部图像检索。 与先前的深度量化方法不同, 即从数据中学习量化的编码词, 我们提出了一个新方案, 使用预先定义的正态矢量作为编码词, 目的是加强量化信息, 并减少代码词的冗余。 为了尽量利用与字典相关的远程测量信息, 我们设计了一个专门化的深度量化方法, 我们设计了一个专门化的分类方法, 目的是为了让所有易读化的代码化能力最大化, 我们设计了一个专门化的模型, 与常规代码化相比, 常规的公式化方法都比常规的高级化方法要更加精确。 与常规化相比, 格式化方法比常规的高级性规则化方法都比高级。