Learning to rank has been intensively studied and widely applied in information retrieval. Typically, a global ranking function is learned from a set of labeled data, which can achieve good performance on average but may be suboptimal for individual queries by ignoring the fact that relevant documents for different queries may have different distributions in the feature space. Inspired by the idea of pseudo relevance feedback where top ranked documents, which we refer as the \textit{local ranking context}, can provide important information about the query's characteristics, we propose to use the inherent feature distributions of the top results to learn a Deep Listwise Context Model that helps us fine tune the initial ranked list. Specifically, we employ a recurrent neural network to sequentially encode the top results using their feature vectors, learn a local context model and use it to re-rank the top results. There are three merits with our model: (1) Our model can capture the local ranking context based on the complex interactions between top results using a deep neural network; (2) Our model can be built upon existing learning-to-rank methods by directly using their extracted feature vectors; (3) Our model is trained with an attention-based loss function, which is more effective and efficient than many existing listwise methods. Experimental results show that the proposed model can significantly improve the state-of-the-art learning to rank methods on benchmark retrieval corpora.
翻译:在信息检索中,已经深入研究并广泛应用了排名学习。通常,从一组标签数据中学习全球排名功能,这些数据可以平均地取得良好的业绩,但对于个别查询来说可能并不理想,因为忽略了不同查询的相关文件在特性空间中分布不同这一事实。受伪相关反馈概念的启发,我们称之为\textit{当地排名背景},它们可以提供与查询特性有关的重要信息。我们提议使用顶级结果的内在特征分布来学习一个有助于我们精细调整初始排名列表的深列表背景模型。具体地说,我们使用一个经常性神经网络来按顺序使用其特性矢量对顶级结果进行编码,学习一个本地背景模型,并使用它重新排列顶级结果。我们的模型有三个优点:(1) 我们的模型可以利用深层神经网络获取基于顶级结果之间复杂互动的本地排名背景环境;(2) 我们的模型可以通过直接使用提取的特征矢量矢量矢量矢量数据来建立现有的学习到排序方法。(3) 我们的模型是经过经常性的神经网络网络网络,用来对顶级结果进行顺序的顺序进行连续的编算,以更高的实验性模型,以显示比现有水平的实验性标准的模型,可以显示较有效的学习方法。