In order to train children's ability to ask curiosity-driven questions, previous research has explored designing specific exercises relying on providing semantic and linguistic cues to help formulate such questions. But despite showing pedagogical efficiency, this method is still limited as it relies on generating the said cues by hand, which can be a very costly process. In this context, we propose to leverage advances in the natural language processing field (NLP) and investigate the efficiency of using a large language model (LLM) for automating the production of the pedagogical content of a curious question-asking (QA) training. We study generating the said content using the "prompt-based" method that consists of explaining the task to the LLM in natural text. We evaluate the output using human experts annotations and comparisons with hand-generated content. Results suggested indeed the relevance and usefulness of this content. We also conduct a field study in primary school (75 children aged 9-10), where we evaluate children's QA performance when having this training. We compare 3 types of content : 1) hand-generated content that proposes "closed" cues leading to predefined questions; 2) GPT-3-generated content that proposes the same type of cues; 3) GPT-3-generated content that proposes "open" cues leading to several possible questions. We see a similar QA performance between the two "closed" trainings (showing the scalability of the approach using GPT-3), and a better one for participants with the "open" training. These results suggest the efficiency of using LLMs to support children in generating more curious questions, using a natural language prompting approach that affords usability by teachers and other users not specialists of AI techniques. Furthermore, results also show that open-ended content may be more suitable for training curious question-asking skills.
翻译:为了培养儿童询问好奇问题的能力,先前的研究探索了如何设计具体练习,依靠提供语义和语言提示来帮助提出这类问题。尽管显示教学效率,但这一方法仍然有限,因为它依靠亲手生成所述提示,这可能是一个非常昂贵的过程。在这方面,我们提议利用自然语言处理领域(NLP)的进步,并调查使用大型语言模型(LLM)来自动制作开放的问询(QA)培训教学内容的效率。我们研究使用“快速基础”方法生成所述内容,该方法包括用自然文本向LLLM解释任务。我们利用人专家的说明和手生成的内容比较,结果表明该内容的相关性和实用性。我们还在小学(75名9-10岁儿童)进行实地研究,在进行这种培训时,我们评估儿童QA教师的快速学习成绩。我们比较了三种类型的内容:1手制作的内容,该内容建议“关闭”引导预定义的G-MLM(P)方法,该方法包括向LM(LM)解释任务;我们利用人手智能说明产出,而GPT-3(G-LA)提供类似内容,该类型也提议使用类似内容。</s>