Neural models with an encoder-decoder framework provide a feasible solution to Question Generation (QG). However, after analyzing the model vocabulary we find that current models (both RNN-based and pre-training based) have more than 23\% inflected forms. As a result, the encoder will generate separate embeddings for the inflected forms, leading to a waste of training data and parameters. Even worse, in decoding these models are vulnerable to irrelevant noise and they suffer from high computational costs. In this paper, we propose an approach to enhance the performance of QG by fusing word transformation. Firstly, we identify the inflected forms of words from the input of encoder, and replace them with the root words, letting the encoder pay more attention to the repetitive root words. Secondly, we propose to adapt QG as a combination of the following actions in the encode-decoder framework: generating a question word, copying a word from the source sequence or generating a word transformation type. Such extension can greatly decrease the size of predicted words in the decoder as well as noise. We apply our approach to a typical RNN-based model and \textsc{UniLM} to get the improved versions. We conduct extensive experiments on SQuAD and MS MARCO datasets. The experimental results show that the improved versions can significantly outperform the corresponding baselines in terms of BLEU, ROUGE-L and METEOR as well as time cost.
翻译:带有编码器- decoder 框架的神经模型为问题生成提供了可行的解决方案。 然而,在分析了模型词汇之后,我们发现目前的模型(基于 RNN 和前培训基础)有23 ⁇ 以上反射形式。因此,编码器将为隐含形式产生单独的嵌入,从而导致培训数据和参数的浪费。更糟糕的是,在解码这些模型时很容易受到不相关的噪音的影响,而且它们会受到高计算成本的影响。在本文中,我们提出一种方法,通过使用文字转换来提高QG的性能。首先,我们从编码器输入的输入中找出隐含的单词形式,并用根字取代它们。因此,编码器将产生不同的嵌入式嵌入,从而导致培训数据和参数的浪费。我们提议将QG作为编码解码-decoder框架中以下行动的组合:生成一个问题单词,复制源序列中的单词,或生成一个词变变换类型。这种扩展可以大大降低REG的模型和变式的模型的大小,将S-NUL 格式作为典型的S-RODRM 的模型,并显示我们的S-RADR-L 改进的数据版本。