Educators are rapidly switching to remote proctoring and examination software for their testing needs, both due to the COVID-19 pandemic and the expanding virtualization of the education sector. State boards are increasingly utilizing these software for high stakes legal and medical licensing exams. Three key concerns arise with the use of these complex software: exam integrity, exam procedural fairness, and exam-taker security and privacy. We conduct the first technical analysis of each of these concerns through a case study of four primary proctoring suites used in U.S. law school and state attorney licensing exams. We reverse engineer these proctoring suites and find that despite promises of high-security, all their anti-cheating measures can be trivially bypassed and can pose significant user security risks. We evaluate current facial recognition classifiers alongside the classifier used by Examplify, the legal exam proctoring suite with the largest market share, to ascertain their accuracy and determine whether faces with certain skin tones are more readily flagged for cheating. Finally, we offer recommendations to improve the integrity and fairness of the remotely proctored exam experience.
翻译:由于COVID-19大流行以及教育部门日益虚拟化,教育工作者正在迅速转向远程测试和考试软件,以满足其测试需要。国家董事会正在越来越多地利用这些软件进行高风险法律和医疗许可证考试。使用这些复杂的软件产生了三个关键问题:考试完整性、程序公平性、考试教师安全和隐私。我们通过对美国法学院和州检察官许可证考试中使用的四套初级测试套房进行案例研究,对其中每一个问题进行首次技术分析。我们对这些测试套房进行改造,发现尽管有高度安全的承诺,但是它们的所有抗切片措施都可能被轻描淡写地绕过,并可能给用户带来重大的安全风险。我们评估目前与Examplify使用的分类器一道的面部识别分类器,即拥有最大市场份额的法律测试测试套房,以确定其准确性,并确定与某些皮肤的面部是否更容易被标记作欺骗。最后,我们提出建议,以提高遥控测试经验的完整性和公正性。