An important task for designing QA systems is answer sentence selection (AS2): selecting the sentence containing (or constituting) the answer to a question from a set of retrieved relevant documents. In this paper, we propose three novel sentence-level transformer pre-training objectives that incorporate paragraph-level semantics within and across documents, to improve the performance of transformers for AS2, and mitigate the requirement of large labeled datasets. Specifically, the model is tasked to predict whether: (i) two sentences are extracted from the same paragraph, (ii) a given sentence is extracted from a given paragraph, and (iii) two paragraphs are extracted from the same document. Our experiments on three public and one industrial AS2 datasets demonstrate the empirical superiority of our pre-trained transformers over baseline models such as RoBERTa and ELECTRA for AS2.
翻译:设计质量保证系统的一项重要任务是回答选择(AS2):从一套检索到的相关文件中选择含有(或构成)对一个问题的答复的句子;在本文件中,我们提出三个新的句级变压器培训前新目标,将文件内部和跨文档的段落级语义纳入其中,以改进AS2变压器的性能,并减轻对大标签数据集的要求;具体地说,模型的任务是预测:(一) 从同一段抽取两个句子;(二) 从某一段抽取一个给定的句子;(三) 从同一文件中提取两个段落;我们对三个公共和一个工业变压器数据集的实验表明,我们预先培训的变压器在经验上优于RABERTA和AS2ELECTRA等基线模型。