We address the problem of incremental sequence classification, where predictions are updated as new elements in the sequence are revealed. Drawing on temporal-difference learning from reinforcement learning, we identify a temporal-consistency condition that successive predictions should satisfy. We leverage this condition to develop a novel loss function for training incremental sequence classifiers. Through a concrete example, we demonstrate that optimizing this loss can offer substantial gains in data efficiency. We apply our method to text classification tasks and show that it improves predictive accuracy over competing approaches on several benchmark datasets. We further evaluate our approach on the task of verifying large language model generations for correctness in grade-school math problems. Our results show that models trained with our method are better able to distinguish promising generations from unpromising ones after observing only a few tokens.
翻译:我们研究增量序列分类问题,即随着序列中新元素的逐步揭示而持续更新预测结果。借鉴强化学习中的时序差分学习思想,我们提出连续预测应满足的时间一致性条件。基于此条件,我们开发了一种用于训练增量序列分类器的新型损失函数。通过具体案例,我们证明优化该损失函数能显著提升数据利用效率。我们将本方法应用于文本分类任务,在多个基准数据集上的实验表明,其预测准确性优于现有竞争方法。此外,我们在验证大型语言模型生成小学数学习题解答的正确性任务中进一步评估了本方法。实验结果表明,采用本方法训练的模型仅需观测少量词元即可更有效地区分有潜力的生成结果与无潜力的生成结果。