Online continual learning (OCL) aims to enable model learning from a non-stationary data stream to continuously acquire new knowledge as well as retain the learnt one, under the constraints of having limited system size and computational cost, in which the main challenge comes from the "catastrophic forgetting" issue -- the inability to well remember the learnt knowledge while learning the new ones. With the specific focus on the class-incremental OCL scenario, i.e. OCL for classification, the recent advance incorporates the contrastive learning technique for learning more generalised feature representation to achieve the state-of-the-art performance but is still unable to fully resolve the catastrophic forgetting. In this paper, we follow the strategy of adopting contrastive learning but further introduce the semantically distinct augmentation technique, in which it leverages strong augmentation to generate more data samples, and we show that considering these samples semantically different from their original classes (thus being related to the out-of-distribution samples) in the contrastive learning mechanism contributes to alleviate forgetting and facilitate model stability. Moreover, in addition to contrastive learning, the typical classification mechanism and objective (i.e. softmax classifier and cross-entropy loss) are included in our model design for faster convergence and utilising the label information, but particularly equipped with a sampling strategy to tackle the tendency of favouring the new classes (i.e. model bias towards the recently learnt classes). Upon conducting extensive experiments on CIFAR-10, CIFAR-100, and Mini-Imagenet datasets, our proposed method is shown to achieve superior performance against various baselines.
翻译:在线持续学习(OCL)的目标是使模型从非静止数据流中学习能够不断获得新知识,并保留在系统规模和计算成本有限的限制下,在系统规模和计算成本有限的限制下,使模型学习能够不断获得新知识并保留所学的知识,主要挑战来自“灾难性遗忘”问题 -- -- 无法在学习新数据的同时很好地记住所学知识。由于具体地侧重于类级增长的OCL情景,即用于分类的OCL,最近的进展包括了对比性学习更概括性特征的学习技术,以获得最新业绩,但仍无法完全解决灾难性的遗忘问题。在本文件中,我们遵循了采用对比性学习的战略,但又进一步引入了具有分层差异的增强技术,即它利用强大的增强力生成更多的数据样本。我们表明,在对比性学习机制中,这些样本与原类别(因此与分配外样本有关)不同,最近的进展是减轻记忆并促进模型稳定性。 此外,除了对比性学习之外,典型的分类机制以及最近在模拟性研究中显示的、特别是模拟性、模拟性、模拟性、模拟性、模拟性、模拟性、模拟性、模拟性、模拟式、模拟式、模拟式、模拟式、模拟式、模拟、模拟式、模拟式、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、模拟、