Although existing machine reading comprehension models are making rapid progress on many datasets, they are far from robust. In this paper, we propose an understanding-oriented machine reading comprehension model to address three kinds of robustness issues, which are over sensitivity, over stability and generalization. Specifically, we first use a natural language inference module to help the model understand the accurate semantic meanings of input questions so as to address the issues of over sensitivity and over stability. Then in the machine reading comprehension module, we propose a memory-guided multi-head attention method that can further well understand the semantic meanings of input questions and passages. Third, we propose a multilanguage learning mechanism to address the issue of generalization. Finally, these modules are integrated with a multi-task learning based method. We evaluate our model on three benchmark datasets that are designed to measure models robustness, including DuReader (robust) and two SQuAD-related datasets. Extensive experiments show that our model can well address the mentioned three kinds of robustness issues. And it achieves much better results than the compared state-of-the-art models on all these datasets under different evaluation metrics, even under some extreme and unfair evaluations. The source code of our work is available at: https://github.com/neukg/RobustMRC.
翻译:虽然现有的机器阅读理解模型在许多数据集上正在取得快速进展,但它们远非稳健。 在本文中,我们提议了一个面向理解的机器阅读理解模型,以解决三种类型的稳健性问题,这些问题涉及敏感度、稳定性和概括性。具体地说,我们首先使用自然语言推断模块,帮助模型理解输入问题的准确语义含义,以便解决过敏度和稳定性的问题。然后在机器阅读理解模块中,我们提议了一个记忆引导多头关注方法,可以进一步理解投入问题和段落的语义含义。第三,我们提议了一个多语言学习机制,以解决笼统化问题。最后,这些模块与多任务学习法相结合。我们评估了三个基准数据集的模型,目的是测量模型的稳健性,包括DuReader(robust)和两个SQUAD相关数据集。广泛的实验显示,我们的模型可以很好地解决上述三种稳健问题。在比较的州/州/州/州/州/州标准下,它取得了更好的结果。这些模块与多任务学习法的模型是用来测量模型下所有数据源的不公平的。