Today the pre-trained language models achieve great success for question generation (QG) task and significantly outperform traditional sequence-to-sequence approaches. However, the pre-trained models treat the input passage as a flat sequence and are thus not aware of the text structure of input passage. For QG task, we model text structure as answer position and syntactic dependency, and propose answer localness modeling and syntactic mask attention to address these limitations. Specially, we present localness modeling with a Gaussian bias to enable the model to focus on answer-surrounded context, and propose a mask attention mechanism to make the syntactic structure of input passage accessible in question generation process. Experiments on SQuAD dataset show that our proposed two modules improve performance over the strong pre-trained model ProphetNet, and combing them together achieves very competitive results with the state-of-the-art pre-trained model.
翻译:今天,经过培训的语言模式在问题生成任务中取得了巨大成功,并大大超越了传统的顺序顺序方法。然而,经过培训的模式将输入通道视为一个平坦的顺序,因此不了解输入通道的文字结构。对于 QG 任务,我们将文本结构建模为回答位置和同步依赖,并提议对回答位置进行建模和合成掩码,以克服这些限制。 特别是,我们用高斯式的偏差模拟当地特征,使模型能够侧重于回答环绕背景,并提议一个掩盖关注机制,使输入通道的合成结构在问题生成过程中可以进入。 SQUAD 数据集实验显示,我们提出的两个模块改进了经过培训的强大模型先知网的性能,并结合了经过培训的州级模型,取得了非常有竞争力的结果。