Class incremental learning (CIL) is the process of continually learning new object classes from incremental data while not forgetting past learned classes. While the common method for evaluating CIL algorithms is based on average test accuracy for all learned classes, we argue that maximizing accuracy alone does not necessarily lead to effective CIL algorithms. In this paper, we experimentally analyze neural network models trained by CIL algorithms using various evaluation protocols in representation learning and propose a new analysis method. Our experiments show that most state-of-the-art algorithms prioritize high stability and do not significantly change the learned representation, and sometimes even learn a representation of lower quality than a naive baseline. However, we observe that these algorithms can still achieve high test accuracy because they learn a classifier that is closer to the optimal classifier. We also found that the base model learned in the first task varies in representation quality across different algorithms, and changes in the final performance were observed when each algorithm was trained under similar representation quality of the base model. Thus, we suggest that representation-level evaluation is an additional recipe for more objective evaluation and effective development of CIL algorithms.
翻译:渐进式学习(CIL)是指从增量数据中不断学习新的对象类别而不会忘记过去所学类别的过程。虽然评估CIL算法的常见方法是基于所有已学习类别的平均测试准确率,但我们认为仅仅最大化准确率并不能一定导致有效的CIL算法。在本文中,我们利用基于表示学习的各种评估方法,对以CIL算法训练的神经网络模型进行实验分析并提出一种新的分析方法。我们的实验表明,大多数最先进的算法重视高稳定性并且不会显著改变所学表示,有时甚至学习到比一个简单的基准线更差的表示。然而,我们发现这些算法仍然可以达到高测试准确率,因为它们学习了一个更接近最优分类器的分类器。此外,我们还发现,第一个任务学习的基本模型在不同算法中所学表示的质量会有所不同,并且当每个算法在相似的基本模型表示质量下进行训练时,最终性能会发生变化。因此,我们建议通过表示级别评估来更客观地评估和有效地开发CIL算法。