Throughout schooling, students are tested on reading comprehension and logical reasoning. Students have developed various strategies for completing such exams, some of which are generally thought to outperform others. One such strategy involves emphasizing relative accuracy over absolute accuracy and can theoretically produce the correct answer without full knowledge of the information required to solve the question. This paper examines the effectiveness of applying such a strategy to train transfer learning models to solve reading comprehension and logical reasoning questions. The models were evaluated on the ReClor dataset, a challenging reading comprehension and logical reasoning benchmark. While previous studies targeted logical reasoning skills, we focus on a general training method and model architecture. We propose the polytuplet loss function, an extension of the triplet loss function, to ensure prioritization of learning the relative correctness of answer choices over learning the true accuracy of each choice. Our results indicate that models employing polytuplet loss outperform existing baseline models. Although polytuplet loss is a promising alternative to other contrastive loss functions, further research is required to quantify the benefits it may present.
翻译:在整个学校教育过程中,学生会接受阅读理解和逻辑推理方面的考试。学生们已经制定了各种完成此类考试的策略,其中一些被普遍认为比其他策略表现更好。其中一种策略是强调相对准确性而不是绝对准确性的策略,理论上可以在不完全掌握所需信息的情况下产生正确答案。本文研究了将这种策略应用于训练迁移学习模型以解决阅读理解和逻辑推理问题的有效性。模型在ReClor数据集上进行了评估,这是一个具有挑战性的阅读理解和逻辑推理基准。虽然以前的研究针对逻辑推理技能,但我们专注于一种通用的训练方法和模型架构。我们提出了多元组损失函数,这是三元组损失函数的扩展,以确保重点学习答案选项相对正确性而不是每个选项的真实准确性。我们的结果表明,采用多元组损失的模型优于现有基准模型。虽然多元组损失是其他对比损失函数的有希望的替代方案,但需要进一步研究以量化其可能带来的好处。