Recent work has shown that commonly available machine reading comprehension (MRC) datasets can be used to train high-performance neural information retrieval (IR) systems. However, the evaluation of neural IR has so far been limited to standard supervised learning settings, where they have outperformed traditional term matching baselines. We conduct in-domain and out-of-domain evaluations of neural IR, and seek to improve its robustness across different scenarios, including zero-shot settings. We show that synthetic training examples generated using a sequence-to-sequence generator can be effective towards this goal: in our experiments, pre-training with synthetic examples improves retrieval performance in both in-domain and out-of-domain evaluation on five different test sets.
翻译:最近的工作表明,现有通用机读理解数据集可用于培训高性能神经信息检索系统,然而,迄今为止,神经IR的评估仅限于标准监督的学习环境,其绩效超过了传统的匹配基线。 我们对神经IR进行内部和外部评估,并设法提高它在不同情景中的稳健性,包括零射场。我们表明,使用序列到序列生成器生成的合成培训实例可以有效地实现这一目标:在我们的实验中,以合成示例进行的培训前提高了在五套不同的测试中在内部和场外评估的检索性能。