Pre-trained Language Models (PLMs) have achieved great success on Machine Reading Comprehension (MRC) over the past few years. Although the general language representation learned from large-scale corpora does benefit MRC, the poor support in evidence extraction which requires reasoning across multiple sentences hinders PLMs from further advancing MRC. To bridge the gap between general PLMs and MRC, we present REPT, a REtrieval-based Pre-Training approach. In particular, we introduce two self-supervised tasks to strengthen evidence extraction during pre-training, which is further inherited by downstream MRC tasks through the consistent retrieval operation and model architecture. To evaluate our proposed method, we conduct extensive experiments on five MRC datasets that require collecting evidence from and reasoning across multiple sentences. Experimental results demonstrate the effectiveness of our pre-training approach. Moreover, further analysis shows that our approach is able to enhance the capacity of evidence extraction without explicit supervision.
翻译:过去几年来,经过培训的语文模式在机器阅读理解(MRC)方面取得了巨大成功,尽管从大型公司中学到的一般语言代表性确实有益于MRC,但在证据提取方面缺乏支持,需要跨多个句子进行推理,这阻碍了PLM进一步推进MRC。为了缩小一般PLM和MRC之间的差距,我们提出了基于RETREVAL的训练前方法REPT。特别是,我们引入了两项由自我监督的任务,在培训前加强证据提取,这是由下游MRC的任务通过连贯的检索操作和模型结构进一步继承的。为了评估我们拟议的方法,我们对需要从多个句子收集证据和推理的五套MRC数据集进行了广泛的实验。实验结果表明我们的培训前方法的有效性。此外,进一步的分析表明,我们的方法能够在没有明确监督的情况下提高证据提取能力。