Text retrieval using learned dense representations has recently emerged as a promising alternative to "traditional" text retrieval using sparse bag-of-words representations. One recent work that has garnered much attention is the dense passage retriever (DPR) technique proposed by Karpukhin et al. (2020) for end-to-end open-domain question answering. We present a replication study of this work, starting with model checkpoints provided by the authors, but otherwise from an independent implementation in our group's Pyserini IR toolkit and PyGaggle neural text ranking library. Although our experimental results largely verify the claims of the original paper, we arrived at two important additional findings that contribute to a better understanding of DPR: First, it appears that the original authors under-report the effectiveness of the BM25 baseline and hence also dense--sparse hybrid retrieval results. Second, by incorporating evidence from the retriever and an improved answer span scoring technique, we are able to improve end-to-end question answering effectiveness using exactly the same models as in the original work.
翻译:最近一项引起人们极大关注的工作是Karpukhin等人(2020年)为端至端开放域问题解答提议的密集通道检索技术。我们介绍了对这项工作的复制研究,首先从作者提供的示范检查站开始,但从本组的Pyserini IR工具包和PyGagle神经文字排名库的独立实施开始,我们通过实验结果基本上证实了原始文件的主张,我们又得出了两项重要结论,有助于更好地了解DPR:首先,原始作者似乎在报告BM25基线的有效性,从而也发现了密集的混合检索结果。第二,通过纳入检索者提供的证据,改进了回答评分技术,我们得以利用与原始工作完全相同的模型改进了端对端回答的有效性。