Educational technologies nowadays increasingly use data and Machine Learning (ML) models. This gives the students, instructors, and administrators support and insights for the optimum policy. However, it is well acknowledged that ML models are subject to bias, which raises concerns about the fairness, bias, and discrimination of using these automated ML algorithms in education and its unintended and unforeseen negative consequences. The contribution of bias during the decision-making comes from datasets used for training ML models and the model architecture. This paper presents a preliminary investigation of the fairness of transformer neural networks on the two tabular datasets: Law School and Student-Mathematics. In contrast to classical ML models, the transformer-based models transform these tabular datasets into a richer representation while solving the classification task. We use different fairness metrics for evaluation and check the trade-off between fairness and accuracy of the transformer-based models over the tabular datasets. Empirically, our approach shows impressive results regarding the trade-off between fairness and performance on the Law School dataset.
翻译:目前,教育技术越来越多地使用数据和机器学习模式。这为学生、教官和行政人员提供了最佳政策的支持和见解。然而,人们公认,ML模式受到偏见的影响,这引起了人们对在教育中使用这些自动ML算法的公平、偏见和歧视及其意外和意外的负面后果的关切。在决策过程中,偏见的推波助澜来自用于培训ML模型和模型结构的数据集。本文件对两个表格数据集(法学院和学生数学)变异神经网络的公平性进行了初步调查。与传统的ML模型不同,基于变异器的模型将这些表格数据集转换成更富的代表性,同时解决分类任务。我们使用不同的公平性衡量标准进行评估,并检查基于变异模型在表格数据集上的公平和准确性之间的权衡。我们的方法很生动地表明,在法学院数据集的公平与业绩之间的权衡方面,取得了令人印象深刻的结果。