The advent of deep neural networks pre-trained via language modeling tasks has spurred a number of successful applications in natural language processing. This work explores one such popular model, BERT, in the context of document ranking. We propose two variants, called monoBERT and duoBERT, that formulate the ranking problem as pointwise and pairwise classification, respectively. These two models are arranged in a multi-stage ranking architecture to form an end-to-end search system. One major advantage of this design is the ability to trade off quality against latency by controlling the admission of candidates into each pipeline stage, and by doing so, we are able to find operating points that offer a good balance between these two competing metrics. On two large-scale datasets, MS MARCO and TREC CAR, experiments show that our model produces results that are either at or comparable to the state of the art. Ablation studies show the contributions of each component and characterize the latency/quality tradeoff space.
翻译:通过语言建模任务预先培训的深层神经网络的出现,激发了自然语言处理方面的一些成功应用。这项工作探索了这样一种流行模式,即BERT,在文件排名方面探索了这样一个模式。我们提出了两个变体,即PIBERT和duoBERT,分别将排名问题分为点分类和对比分类。这两种模式安排在一个多阶段的排名结构中,形成一个端对端搜索系统。这一设计的一个主要优点是能够通过控制候选人进入每个编审阶段来交换质量和延缓,这样我们就能找到操作点,在这两种相互竞争的指标之间保持良好的平衡。在两个大型数据集(MS MARCO和TREC CAR)上,实验显示,我们的模型产生的结果要么在艺术状态上,要么可以与艺术状态相仿。缩缩图研究表明了每个组成部分的贡献,并说明了纬度/质量交易空间的特点。