Computerized adaptive testing (CAT) refers to a form of tests that are personalized to every student/test taker. CAT methods adaptively select the next most informative question/item for each student given their responses to previous questions, effectively reducing test length. Existing CAT methods use item response theory (IRT) models to relate student ability to their responses to questions and static question selection algorithms designed to reduce the ability estimation error as quickly as possible; therefore, these algorithms cannot improve by learning from large-scale student response data. In this paper, we propose BOBCAT, a Bilevel Optimization-Based framework for CAT to directly learn a data-driven question selection algorithm from training data. BOBCAT is agnostic to the underlying student response model and is computationally efficient during the adaptive testing process. Through extensive experiments on five real-world student response datasets, we show that BOBCAT outperforms existing CAT methods (sometimes significantly) at reducing test length.
翻译:计算机化适应性测试(CAT)是指针对每个学生/测试者的个人测试形式。 CAT方法根据每个学生对先前问题的答复,适应性地为每个学生选择下一个信息最丰富的问题/项目,有效地缩短了测试长度。 现有的CAT方法使用项目反应理论模型,将学生的能力与其对问题的答复和静态问题选择算法联系起来,这些算法旨在尽快减少能力估计错误;因此,这些算法无法通过学习大规模学生反应数据来改进。 在本文中,我们建议计算机化技术双级优化框架BOBCAT, 直接从培训数据中学习数据驱动的问题选择算法。 BOBCAT对基础学生反应模型具有不可知性,在适应性测试过程中具有计算效率。 通过对五个现实世界学生反应数据集的广泛实验,我们显示BOBCAT在缩短测试长度时比现有的计算机化方法(有时显著)要强。