Ranking has always been one of the top concerns in information retrieval researches. For decades, the lexical matching signal has dominated the ad-hoc retrieval process, but solely using this signal in retrieval may cause the vocabulary mismatch problem. In recent years, with the development of representation learning techniques, many researchers turn to Dense Retrieval (DR) models for better ranking performance. Although several existing DR models have already obtained promising results, their performance improvement heavily relies on the sampling of training examples. Many effective sampling strategies are not efficient enough for practical usage, and for most of them, there still lacks theoretical analysis in how and why performance improvement happens. To shed light on these research questions, we theoretically investigate different training strategies for DR models and try to explain why hard negative sampling performs better than random sampling. Through the analysis, we also find that there are many potential risks in static hard negative sampling, which is employed by many existing training methods. Therefore, we propose two training strategies named a Stable Training Algorithm for dense Retrieval (STAR) and a query-side training Algorithm for Directly Optimizing Ranking pErformance (ADORE), respectively. STAR improves the stability of DR training process by introducing random negatives. ADORE replaces the widely-adopted static hard negative sampling method with a dynamic one to directly optimize the ranking performance. Experimental results on two publicly available retrieval benchmark datasets show that either strategy gains significant improvements over existing competitive baselines and a combination of them leads to the best performance.
翻译:在信息检索研究中,排名一直是最令人关切的问题之一。几十年来,词汇匹配信号一直主导着临时检索过程,但只是在检索中仅使用这一信号可能会造成词汇错配问题。近年来,随着代表性学习技术的发展,许多研究人员转向Dense Retreival(DR)模型,以取得更好的排名业绩。虽然现有的几个DR模型已经取得了有希望的结果,但其性能改进在很大程度上依赖于培训实例的抽样。许多有效的抽样战略对于实际使用来说效率不够高,而且对于大多数这类模式来说,在改进绩效的方式和原因方面仍然缺乏理论分析。为了揭示这些研究问题,我们理论上调查不同的DR模型培训战略,并试图解释为什么硬性抽样比随机抽样效果更好。通过分析,我们还发现在静态的硬性抽样中存在许多潜在风险,而许多现有的培训方法都采用了这种模式。因此,我们建议采用两种培训战略,称为“稳定培训Algorith”法,用于密集的优化再定位(STAR),对于如何改进业绩改进绩效的理论性能直接优化的精确度,或者直接将SDREththimal Stregres Stabistring Stabidustring Stabidustring Stabidustral 的成绩展示,分别分别以公开地显示一种稳定。