Continual learning -- accumulating knowledge from a sequence of learning experiences -- is an important yet challenging problem. In this paradigm, the model's performance for previously encountered instances may substantially drop as additional data are seen. When dealing with class-imbalanced data, forgetting is further exacerbated. Prior work has proposed replay-based approaches which aim at reducing forgetting by intelligently storing instances for future replay. Although Class-Balancing Reservoir Sampling (CBRS) has been successful in dealing with imbalanced data, the intra-class diversity has not been accounted for, implicitly assuming that each instance of a class is equally informative. We present Diverse-CBRS (D-CBRS), an algorithm that allows us to consider within class diversity when storing instances in the memory. Our results show that D-CBRS outperforms state-of-the-art memory management continual learning algorithms on data sets with considerable intra-class diversity.
翻译:持续学习 -- -- 从一系列学习经验中积累知识 -- -- 是一个重要但具有挑战性的问题。在这个范例中,模型在以往遇到的事例中的性能会随着额外数据的出现而大幅下降。在处理课堂平衡数据时,遗忘现象会进一步恶化。先前的工作提出了以重放为基础的方法,旨在通过明智地储存案例来减少遗忘,以便将来重播。虽然级平衡回收储量抽样(CBRS)在处理不平衡的数据方面很成功,但是没有考虑到阶级内部的多样性,这意味着每个班级都有相同的信息。我们提出了多元-CBRS(D-CBRS)算法,允许我们在存储记忆中的事件时在课堂内考虑多样性。我们的结果显示,D-CBRS(D-CBRS)在数据集上超越了最新的记忆管理持续学习算法,而该数据集具有相当大的阶级内部多样性。