A primary goal of class-incremental learning is to strike a balance between stability and plasticity, where models should be both stable enough to retain knowledge learned from previously seen classes, and plastic enough to learn concepts from new classes. While previous works demonstrate strong performance on class-incremental benchmarks, it is not clear whether their success comes from the models being stable, plastic, or a mixture of both. This paper aims to shed light on how effectively recent class-incremental learning algorithms address the stability-plasticity trade-off. We establish analytical tools that measure the stability and plasticity of feature representations, and employ such tools to investigate models trained with various algorithms on large-scale class-incremental benchmarks. Surprisingly, we find that the majority of class-incremental learning algorithms heavily favor stability over plasticity, to the extent that the feature extractor of a model trained on the initial set of classes is no less effective than that of the final incremental model. Our observations not only inspire two simple algorithms that highlight the importance of feature representation analysis, but also suggest that class-incremental learning approaches, in general, should strive for better feature representation learning.
翻译:增量学习的一个主要目标是在稳定性和可塑性之间达到平衡,既要保持模型稳定以保留先前学习过的类别的知识,又要兼具可塑性以学习新类别的概念。虽然之前的工作展示了在增量类别基准测试上强大的性能,但并不清楚它们的成功是否来自于模型的稳定性、可塑性,还是两者的混合。本文旨在阐明最近的增量学习算法如何有效地解决稳定性-可塑性交换问题。我们建立了衡量特征表示稳定性和可塑性的分析工具,并使用这些工具调查在大规模增量基准上使用各种算法训练的模型。令人惊讶的是,我们发现大多数增量学习算法在稳定性的基础上重视可塑性,以至于训练在初始类别集上的模型的特征提取器与最终增量模型的特征提取器一样有效。我们的观察不仅启发了两种简单的算法,凸显了特征表示分析的重要性,而且还表明增量学习方法通常应努力改进特征表示学习。