Recent years have witnessed the significant advance in dense retrieval (DR) based on powerful pre-trained language models (PLM). DR models have achieved excellent performance in several benchmark datasets, while they are shown to be not as competitive as traditional sparse retrieval models (e.g., BM25) in a zero-shot retrieval setting. However, in the related literature, there still lacks a detailed and comprehensive study on zero-shot retrieval. In this paper, we present the first thorough examination of the zero-shot capability of DR models. We aim to identify the key factors and analyze how they affect zero-shot retrieval performance. In particular, we discuss the effect of several key factors related to source training set, analyze the potential bias from the target dataset, and review and compare existing zero-shot DR models. Our findings provide important evidence to better understand and develop zero-shot DR models.
翻译:近年来,在强大的预先培训语言模型(PLM)基础上的密集检索(DR)取得了显著进展。 DR模型在若干基准数据集中取得了优异的成绩,而事实证明,在零发检索环境中,这些模型的竞争力不如传统的稀疏检索模型(例如BM25)那么强,然而,在相关文献中,仍然缺乏关于零发检索的详细和全面的研究。在本文件中,我们首次对DR模型的零发能力进行了彻底审查。我们的目标是确定关键因素,分析其对零发检索性能的影响。我们特别讨论了与源培训组有关的几个关键因素的影响,分析了目标数据集的潜在偏差,并审查和比较了现有的零发DR模型。我们的调查结果为更好地理解和发展零发记录模型提供了重要证据。