Machine learning models for the ad-hoc retrieval of documents and passages have recently shown impressive improvements due to better language understanding using large pre-trained language models. However, these over-parameterized models are inherently non-interpretable and do not provide any information on the parts of the documents that were used to arrive at a certain prediction. In this paper we introduce the select and rank paradigm for document ranking, where interpretability is explicitly ensured when scoring longer documents. Specifically, we first select sentences in a document based on the input query and then predict the query-document score based only on the selected sentences, acting as an explanation. We treat sentence selection as a latent variable trained jointly with the ranker from the final output. We conduct extensive experiments to demonstrate that our inherently interpretable select-and-rank approach is competitive in comparison to other state-of-the-art methods and sometimes even outperforms them. This is due to our novel end-to-end training approach based on weighted reservoir sampling that manages to train the selector despite the stochastic sentence selection. We also show that our sentence selection approach can be used to provide explanations for models that operate on only parts of the document, such as BERT.
翻译:最近,由于使用经过预先培训的大型语言模型对语言有更好的理解,对文档和段落进行临时检索的机器学习模式显示出了令人印象深刻的改进。然而,这些超分参数模型本质上是不可解释的,没有提供关于用于作出某种预测的文档部分的任何信息。在本文中,我们引入了文件排序的选择和排名模式,在评分较长的文档时明确确保了可解释性。具体地说,我们首先根据输入查询选择的文档选择句子,然后预测仅根据选定句子的查询文件评分,作为解释。我们把选择句子作为与最后输出的排位员共同培训的潜在变量对待。我们进行了广泛的实验,以证明我们固有的可解释的选位法与其他最先进的方法相比是竞争性的,有时甚至超越了这些方法。这要归功于我们基于加权储油层抽样的新式的端到端培训方法,尽管选择了精选的句子,但我们的评分方法也表明,我们的评分方法可以用来为仅使用文件部分的模型提供解释,作为B的模型。