A popular recent approach to answering open-domain questions is to first search for question-related passages and then apply reading comprehension models to extract answers. Existing methods usually extract answers from single passages independently. But some questions require a combination of evidence from across different sources to answer correctly. In this paper, we propose two models which make use of multiple passages to generate their answers. Both use an answer-reranking approach which reorders the answer candidates generated by an existing state-of-the-art QA model. We propose two methods, namely, strength-based re-ranking and coverage-based re-ranking, to make use of the aggregated evidence from different passages to better determine the answer. Our models have achieved state-of-the-art results on three public open-domain QA datasets: Quasar-T, SearchQA and the open-domain version of TriviaQA, with about 8 percentage points of improvement over the former two datasets.
翻译:回答开放领域问题的最新流行方法是首先寻找与问题有关的段落,然后运用阅读理解模型来获取答案。现有方法通常独立地从单个段落中提取答案。但有些问题需要从不同来源收集各种证据才能正确回答。在本文件中,我们提出了两种模式,利用多个段落来生成答案。两者都采用对回答进行重新排序的方法,对现有最新QA模型生成的回答候选人进行重新排序。我们提出了两种方法,即基于实力的重新排序和基于覆盖面的重新排序,以利用不同段落的汇总证据更好地确定答案。我们的模型在三个公开开放领域QA数据集(Quasar-T、SearchQA和TriviaQA开放域版本)上取得了最新的结果,前两个数据集的改进率约为8个百分点。