We present in this work a fully Transformer-based reinforcement learning generator-evaluator architecture for neural question generation. Question generation is a task that consists in generating questions given a context and answer. To improve the quality of the generated question, we came up with a semantic-based self-critical training layout in generator-evaluator architecture, which goes beyond typical maximum likelihood training. Evaluation metrics for language modeling only based on n-gram overlapping do not consider semantic relations between reference and candidate strings. To improve the evaluation step, we assess our model for both n-gram overlap using BLEU and semantically using BERTScore and NUBIA, a novel state-of-the-art evaluation metric for text generation. Question generation could be used in many downstream applications, including in extending question answering datasets, conversational systems, and educational assessment systems.
翻译:在这项工作中,我们为神经问题生成提供了完全基于变换的强化学习、发电机和蒸发器结构。问题生成是一项任务,包括提出有背景和答案的问题。为了提高所产生问题的质量,我们在发电机和蒸发器结构中提出了基于语义的自我批评培训布局,这超出了典型的最大可能性培训范围。仅以正克重叠为基础的语言建模评价指标不考虑参考和候选字符串之间的语义关系。为了改进评估步骤,我们利用BLEU和语言学评估我们的正克重叠模式,使用BERTScore和NUBIA,这是用于文本生成的新颖的、最先进的评价指标。问题生成可用于许多下游应用,包括扩展回答问题的数据集、对话系统和教育评估系统。