It has been observed that neural networks perform poorly when the data or tasks are presented sequentially. Unlike humans, neural networks suffer greatly from catastrophic forgetting, making it impossible to perform life-long learning. To address this issue, memory-based continual learning has been actively studied and stands out as one of the best-performing methods. We examine memory-based continual learning and identify that large variation in the representation space is crucial for avoiding catastrophic forgetting. Motivated by this, we propose to diversify representations by using two types of perturbations: model-agnostic variation (i.e., the variation is generated without the knowledge of the learned neural network) and model-based variation (i.e., the variation is conditioned on the learned neural network). We demonstrate that enlarging representational variation serves as a general principle to improve continual learning. Finally, we perform empirical studies which demonstrate that our method, as a simple plug-and-play component, can consistently improve a number of memory-based continual learning methods by a large margin.
翻译:人们注意到,当数据或任务按顺序提出时,神经网络表现不佳。与人类不同,神经网络遭受灾难性的遗忘,无法进行终身学习。为了解决这一问题,积极研究了以记忆为基础的持续学习,并将其作为最优秀的方法之一。我们研究了基于记忆的持续学习,并发现代表空间的巨大差异对于避免灾难性的遗忘至关重要。为此,我们提议通过使用两种类型的扰动来使表达形式多样化:模型-认知变异(即变异是在没有了解所学神经网络的情况下产生的)和模型变异(即变异是以所学的神经网络为条件的)。我们证明,扩大代表性变异是改进持续学习的一般原则。最后,我们进行经验研究,表明我们的方法,作为一个简单的插子和玩子,能够以很大的空间不断改进一些基于记忆的不断学习方法。