We study the effects of data size and quality on the performance on Automated Essay Scoring (AES) engines that are designed in accordance with three different paradigms; A frequency and hand-crafted feature-based model, a recurrent neural network model, and a pretrained transformer-based language model that is fine-tuned for classification. We expect that each type of model benefits from the size and the quality of the training data in very different ways. Standard practices for developing training data for AES engines were established with feature-based methods in mind, however, since neural networks are increasingly being considered in a production setting, this work seeks to inform us as to how to establish better training data for neural networks that will be used in production.
翻译:我们研究数据大小和质量对自动测读引擎性能的影响,这些引擎是根据三种不同模式设计的; 频率和手制地物模型,经常性神经网络模型,以及经过精细调整以进行分类的预先训练的变压器语言模型; 我们期望每一种模型都以非常不同的方式从培训数据的规模和质量中获益; 但是,由于神经网络在生产环境中日益受到考虑,因此,我们用基于地物的方法制定了开发AES发动机培训数据的标准做法,这项工作旨在告诉我们如何为将被用于生产中的神经网络建立更好的培训数据。