We present a framework for improving the performance of a wide class of retrieval models at minimal computational cost. It utilizes precomputed document representations extracted by a base dense retrieval method and involves training a model to jointly score a large set of retrieved candidate documents for each query, while potentially transforming on the fly the representation of each document in the context of the other candidates as well as the query itself. When scoring a document representation based on its similarity to a query, the model is thus aware of the representation of its "peer" documents. We show that our approach leads to substantial improvement in retrieval performance over the base method and over scoring candidate documents in isolation from one another, as in a pair-wise training setting. Crucially, unlike term-interaction rerankers based on BERT-like encoders, it incurs a negligible computational overhead on top of any first-stage method at run time, allowing it to be easily combined with any state-of-the-art dense retrieval method. Finally, concurrently considering a set of candidate documents for a given query enables additional valuable capabilities in retrieval, such as score calibration and mitigating societal biases in ranking.
翻译:我们提出了一个框架,以最低计算成本改进一系列广泛的检索模型的性能,它使用一种基础密集检索方法得出的预先计算的文件表述方法,并涉及培训一种模型,以共同得分每个查询的大量检索的候选文件,同时可能自动地改变其他候选人和查询本身对每个文件的表述方式。在根据与查询相似之处评分一个文件表示方式时,该模型因此了解其“同侪”文件的表述方式。我们表明,我们的方法大大改进了检索方法的性能,并导致在相互隔离的情况下对候选文件进行评分,正如在对等培训设置中那样。 关键是,它不同于基于像BERT一样的术语间重新排序,在任何第一阶段方法的顶部,其计算间接费用微不足道,因此可以很容易地与任何最先进的密集检索方法结合起来。最后,我们同时考虑一套特定查询的候选文件,使得在检索方面具有额外的宝贵能力,例如得分校准和减少排名中的社会偏差。