Class-Incremental Learning (CIL) aims to learn a classification model with the number of classes increasing phase-by-phase. An inherent problem in CIL is the stability-plasticity dilemma between the learning of old and new classes, i.e., high-plasticity models easily forget old classes, but high-stability models are weak to learn new classes. We alleviate this issue by proposing a novel network architecture called Adaptive Aggregation Networks (AANets), in which we explicitly build two types of residual blocks at each residual level (taking ResNet as the baseline architecture): a stable block and a plastic block. We aggregate the output feature maps from these two blocks and then feed the results to the next-level blocks. We adapt the aggregation weights in order to balance these two types of blocks, i.e., to balance stability and plasticity, dynamically. We conduct extensive experiments on three CIL benchmarks: CIFAR-100, ImageNet-Subset, and ImageNet, and show that many existing CIL methods can be straightforwardly incorporated into the architecture of AANets to boost their performances.
翻译:高级强化学习(CIL)旨在学习一个分类模式,使各个班级的数量逐步增加。CIL的一个固有问题是,在新旧班级(即高塑性模型很容易忘记旧班级)的学习之间,即高塑性模型容易忘记旧班级,但高稳定性模型较弱,无法学习新班级。我们通过提出一个叫作适应性聚合网络(AANets)的新型网络结构来缓解这一问题,我们在该结构中明确在每个剩余级(以ResNet为基准结构)建立两种类型的剩余区块:一个稳定的区块和一个塑料区块。我们从这两个区块中汇总产出特征图,然后将结果反馈到下一个级区块。我们调整汇总权重,以平衡这两类区块,即稳定性和可塑性。我们就CIL基准(CIFAR-100)、图像网子集和图像网)进行了广泛的实验,并表明许多现有的CIL方法可以直接纳入AANet的架构,以提高其性能。