Dense retrieval (DR) approaches based on powerful pre-trained language models (PLMs) achieved significant advances and have become a key component for modern open-domain question-answering systems. However, they require large amounts of manual annotations to perform competitively, which is infeasible to scale. To address this, a growing body of research works have recently focused on improving DR performance under low-resource scenarios. These works differ in what resources they require for training and employ a diverse set of techniques. Understanding such differences is crucial for choosing the right technique under a specific low-resource scenario. To facilitate this understanding, we provide a thorough structured overview of mainstream techniques for low-resource DR. Based on their required resources, we divide the techniques into three main categories: (1) only documents are needed; (2) documents and questions are needed; and (3) documents and question-answer pairs are needed. For every technique, we introduce its general-form algorithm, highlight the open issues and pros and cons. Promising directions are outlined for future research.
翻译:根据强大的预先培训语言模型(PLMs)的密集检索方法取得了显著进展,已成为现代开放域问答系统的一个关键组成部分,然而,这些方法需要大量手工说明才能竞争性地进行,这是无法扩大规模的。为解决这一问题,越来越多的研究工作最近侧重于在低资源情景下改进DR性能。这些工作在培训资源方面各不相同,采用多种技术。了解这些差异对于在具体的低资源情景下选择正确技术至关重要。为了促进这种理解,我们提供了对低资源DR主流技术的彻底结构化概览。我们根据其所需资源,将技术分为三大类:(1)只需要文件;(2)需要文件和问题;(3)需要文件和问题对口。对于每一种技术,我们都采用通用的算法,强调开放的问题以及赞成和反对。为今后的研究勾画出有希望的方向。