Although considerable efforts have been devoted to transformer-based ranking models for document search, the relevance-efficiency tradeoff remains a critical problem for ad-hoc ranking. To overcome this challenge, this paper presents BECR (BERT-based Composite Re-Ranking), a composite re-ranking scheme that combines deep contextual token interactions and traditional lexical term-matching features. In particular, BECR exploits a token encoding mechanism to decompose the query representations into pre-computable uni-grams and skip-n-grams. By applying token encoding on top of a dual-encoder architecture, BECR separates the attentions between a query and a document while capturing the contextual semantics of a query. In contrast to previous approaches, this framework does not perform expensive BERT computations during online inference. Thus, it is significantly faster, yet still able to achieve high competitiveness in ad-hoc ranking relevance. Finally, an in-depth comparison between BECR and other start-of-the-art neural ranking baselines is described using the TREC datasets, thereby further demonstrating the enhanced relevance and efficiency of BECR.
翻译:尽管在基于变压器的文件搜索排名模式方面已经作出了相当大的努力,但相关性-效率权衡仍然是临时排序的关键问题。为克服这一挑战,本文件介绍了BECR(基于BERT的复合再朗金),这是将深背景象征性互动和传统的词汇术语匹配特征结合起来的综合重新排序办法。特别是,BECR利用象征性编码机制将查询表分解成可预先计算单克和跳N克。通过在双编码结构顶部应用象征性编码,BECR将查询和文件的注意力区分开来,同时捕捉查询的背景语义。与以往的做法不同,这一框架在网上推断中并不进行昂贵的BERT计算。因此,它大大加快了速度,但仍能够在适应性排序相关性方面实现高度的竞争力。最后,使用TREC数据集对BCR和其他开源的神经排序基线进行了深入比较,从而进一步证明了BECRCR的更大相关性和效率。