Models for reading comprehension (RC) commonly restrict their output space to the set of all single contiguous spans from the input, in order to alleviate the learning problem and avoid the need for a model that generates text explicitly. However, forcing an answer to be a single span can be restrictive, and some recent datasets also include multi-span questions, i.e., questions whose answer is a set of non-contiguous spans in the text. Naturally, models that return single spans cannot answer these questions. In this work, we propose a simple architecture for answering multi-span questions by casting the task as a sequence tagging problem, namely, predicting for each input token whether it should be part of the output or not. Our model substantially improves performance on span extraction questions from DROP and Quoref by 9.9 and 5.5 EM points respectively.
翻译:阅读理解(RC)模型通常将其输出空间限制在输入的所有单一毗连宽度范围内,以缓解学习问题,避免需要一种明确生成文本的模型。然而,强迫回答是一个单一的跨度,可能会有限制性,最近的一些数据集还包括多片问题,即答案是文本中一组非毗连的宽度的问题。当然,返回单个宽度的模型无法回答这些问题。在这项工作中,我们提出一个简单的结构,用于回答多片问题,将任务作为一个序列标记问题,即预测每个输入符号是否应该成为输出的一部分。我们的模型极大地提高了从 DROP 和 Quoref 提取问题的绩效,分别是9.9 和 5.5 EM点。