Multiple-choice questions (MCQs) offer the most promising avenue for skill evaluation in the era of virtual education and job recruiting, where traditional performance-based alternatives such as projects and essays have become less viable, and grading resources are constrained. The automated generation of MCQs would allow assessment creation at scale. Recent advances in natural language processing have given rise to many complex question generation methods. However, the few methods that produce deployable results in specific domains require a large amount of domain-specific training data that can be very costly to acquire. Our work provides an initial foray into MCQ generation under high data-acquisition cost scenarios by strategically emphasizing paraphrasing the question context (compared to the task). In addition to maintaining semantic similarity between the question-answer pairs, our pipeline, which we call AGenT Zero, consists of only pre-trained models and requires no fine-tuning, minimizing data acquisition costs for question generation. AGenT Zero successfully outperforms other pre-trained methods in fluency and semantic similarity. Additionally, with some small changes, our assessment pipeline can be generalized to a broader question and answer space, including short answer or fill in the blank questions.
翻译:在虚拟教育和招工时代,多种选择问题(MCQs)是评价技能的最有希望的渠道,在虚拟教育和招工时代,项目和作文等传统的基于业绩的替代方法已变得不那么可行,而且分级资源也受到限制。自动生成MCQ会促成规模评估。自然语言处理的最近进展产生了许多复杂的问题生成方法。然而,在特定领域产生可部署成果的少数方法需要大量特定领域的培训数据,而获得这些数据的成本可能非常昂贵。我们的工作从战略角度强调问题的背景(与任务相比),从而在高数据获取成本设想下为MCQ的一代提供了初步借鉴。除了保持问答对配方之间的语义性相似性外,我们称之为AGenT Zero的管道仅由预先培训的模式组成,不需要微调,将问题生成的数据获取成本降到最低。AGENT Zero成功地超越了其他在流利和语义相似性方面预先培训过的方法。此外,由于一些小的改动,我们的评估管道可以向更空的提问和空格的答案,包括答案。