Teaching logic effectively requires an understanding of the factors which cause logic students to struggle. Formalization exercises, which require the student to produce a formula corresponding to the natural language sentence, are a good candidate for scrutiny since they tap into the students' understanding of various aspects of logic. We correlate the difficulty of formalization exercises predicted by a previously proposed difficulty estimation algorithm with two empirical difficulty measures on the Grade Grinder corpus, which contains student solutions to FOL exercises. We obtain a moderate correlation with both measures, suggesting that the said algorithm indeed taps into important sources of difficulty but leaves a fair amount of variance uncaptured. We conduct an error analysis, closely examining exercises which were misclassified, with the aim of identifying additional sources of difficulty. We identify three additional factors which emerge from the difficulty analysis, namely predicate complexity, pragmatic factors and typicality of the exercises, and discuss the implications of automated difficulty estimation for logic teaching and explainable AI.
翻译:教学逻辑有效要求理解导致逻辑学生挣扎的因素。正规化练习要求学生提出与自然语言句子相对应的公式,由于它们利用学生对逻辑的各个方面的理解,因此是一个很好的审查对象。我们把先前提议的困难估算算法所预测的正规化练习的困难与Grinder Cample上的两项经验困难计量法(其中包括学生对FOL练习的解决方案)联系起来。我们从这两种计量法中取得了适度的关联,表明上述算法确实利用了重要的困难来源,但留下了相当大的差异。我们进行了错误分析,仔细检查了分类错误的练习,目的是找出更多的困难来源。我们找出了从困难分析中产生的另外三个因素,即上游复杂性、务实因素和典型的练习,并讨论了自动化的难度估算对逻辑教学和可解释的AI的影响。