The staple of human intelligence is the capability of acquiring knowledge in a continuous fashion. In stark contrast, Deep Networks forget catastrophically and, for this reason, the sub-field of Class-Incremental Continual Learning fosters methods that learn a sequence of tasks incrementally, blending sequentially-gained knowledge into a comprehensive prediction. This work aims at assessing and overcoming the pitfalls of our previous proposal Dark Experience Replay (DER), a simple and effective approach that combines rehearsal and Knowledge Distillation. Inspired by the way our minds constantly rewrite past recollections and set expectations for the future, we endow our model with the abilities to i) revise its replay memory to welcome novel information regarding past data ii) pave the way for learning yet unseen classes. We show that the application of these strategies leads to remarkable improvements; indeed, the resulting method - termed eXtended-DER (X-DER) - outperforms the state of the art on both standard benchmarks (such as CIFAR-100 and miniImagenet) and a novel one here introduced. To gain a better understanding, we further provide extensive ablation studies that corroborate and extend the findings of our previous research (e.g. the value of Knowledge Distillation and flatter minima in continual learning setups).
翻译:人类情报的主轴是持续获取知识的能力。 形成鲜明对比的是,深网络(Deep Networks) 以灾难性的方式忘记了过去记忆的重写和对未来的期望,因此,我们把模型的外型能力归结于一) 修改它的重现记忆,欢迎关于过去数据的新信息 ii) 为学习但看不见的课程铺平了道路。我们表明,这些战略的应用导致显著的改进;事实上,由此产生的方法――所谓的eXtend-DER(X-DER)――超越了标准基准(如CIFAR-100和MIDIMagenet)和这里介绍的新型标准基准(如CIFAR-100和MIDIMagenet)的艺术现状。为了获得更好的理解,我们进一步提供了以往研究的不断更新和不断更新的研究成果。