Answer Sentence Selection (AS2) is an efficient approach for the design of open-domain Question Answering (QA) systems. In order to achieve low latency, traditional AS2 models score question-answer pairs individually, ignoring any information from the document each potential answer was extracted from. In contrast, more computationally expensive models designed for machine reading comprehension tasks typically receive one or more passages as input, which often results in better accuracy. In this work, we present an approach to efficiently incorporate contextual information in AS2 models. For each answer candidate, we first use unsupervised similarity techniques to extract relevant sentences from its source document, which we then feed into an efficient transformer architecture fine-tuned for AS2. Our best approach, which leverages a multi-way attention architecture to efficiently encode context, improves 6% to 11% over noncontextual state of the art in AS2 with minimal impact on system latency. All experiments in this work were conducted in English.
翻译:答案选择(AS2)是设计开放域问答系统的有效方法。为了实现低延迟,传统的 AS2 模型单独得对答题,忽略了文件中每个潜在答案的任何信息。相反,为机器阅读理解任务设计的更昂贵的计算成本模型通常会收到一个或一个以上段落作为输入,结果往往更准确。在这项工作中,我们提出了一个将背景信息有效纳入AS2 模型的方法。对于每个应答候选人,我们首先使用未经监督的类似技术从其源文档中提取相关句子,然后将其输入一个高效的变压器结构,对AS2进行微调。我们的最佳方法是利用多路关注结构来有效编码环境,将AS2中艺术的非通俗状态提高6%至11%,对系统定位影响最小。这项工作的所有实验都是用英语进行的。