Vector quantization (VQ) based ANN indexes, such as Inverted File System (IVF) and Product Quantization (PQ), have been widely applied to embedding based document retrieval thanks to the competitive time and memory efficiency. Originally, VQ is learned to minimize the reconstruction loss, i.e., the distortions between the original dense embeddings and the reconstructed embeddings after quantization. Unfortunately, such an objective is inconsistent with the goal of selecting ground-truth documents for the input query, which may cause severe loss of retrieval quality. Recent works identify such a defect, and propose to minimize the retrieval loss through contrastive learning. However, these methods intensively rely on queries with ground-truth documents, whose performance is limited by the insufficiency of labeled data. In this paper, we propose Distill-VQ, which unifies the learning of IVF and PQ within a knowledge distillation framework. In Distill-VQ, the dense embeddings are leveraged as "teachers", which predict the query's relevance to the sampled documents. The VQ modules are treated as the "students", which are learned to reproduce the predicted relevance, such that the reconstructed embeddings may fully preserve the retrieval result of the dense embeddings. By doing so, Distill-VQ is able to derive substantial training signals from the massive unlabeled data, which significantly contributes to the retrieval quality. We perform comprehensive explorations for the optimal conduct of knowledge distillation, which may provide useful insights for the learning of VQ based ANN index. We also experimentally show that the labeled data is no longer a necessity for high-quality vector quantization, which indicates Distill-VQ's strong applicability in practice.
翻译:以 VQ 为基础的 VQ, 以尽量减少重建损失。 不幸的是, 此目标与为输入查询选择地真文件的目标不符, 诸如 Inverected File System (IVF) 和产品量化(PQ), 由于竞争性时间和内存效率, 已被广泛用于嵌入基于文件的检索。 最初, VQ 学会了将重建损失降到最低, 即原始密集嵌入和在量化后重建的嵌入之间的扭曲。 不幸的是, 此目标与为输入查询选择地真文件的目标不符, 这可能导致检索质量严重丢失。 最近的工作确定了这样一个缺陷, 并提议通过对比性学习来将检索损失降到最低。 然而, 这些方法非常依赖地真伪文档的查询, 其性因标签数据不足而受到限制。 我们提议, 蒸馏- VQ, 将密度嵌入为“ 教师 ”, 并预测与未取样的文件的相关性。 VQ 模块的精确性, 用于不断复制数据。