Deep Neural Networks (DNNs) suffer from a rapid decrease in performance when trained on a sequence of tasks where only data of the most recent task is available. This phenomenon, known as catastrophic forgetting, prevents DNNs from accumulating knowledge over time. Overcoming catastrophic forgetting and enabling continual learning is of great interest since it would enable the application of DNNs in settings where unrestricted access to all the training data at any time is not always possible, e.g. due to storage limitations or legal issues. While many recently proposed methods for continual learning use some training examples for rehearsal, their performance strongly depends on the number of stored examples. In order to improve performance of rehearsal for continual learning, especially for a small number of stored examples, we propose a novel way of learning a small set of synthetic examples which capture the essence of a complete dataset. Instead of directly learning these synthetic examples, we learn a weighted combination of shared components for each example that enables a significant increase in memory efficiency. We demonstrate the performance of our method on commonly used datasets and compare it to recently proposed related methods and baselines.
翻译:深神经网络(DNNs)在接受仅掌握最新任务数据的任务序列培训时表现迅速下降。 这种现象被称为灾难性忘却,使DNNs无法长期积累知识。 克服灾难性忘却和不断学习是一件非常有意义的事情,因为它将使DNNs能够在有时无法无限制地随时访问所有培训数据的环境中应用,例如由于储存限制或法律问题等原因,有时无法随时不受限制地访问所有培训数据。 虽然许多最近提出的持续学习方法使用一些培训实例进行排练,但其绩效在很大程度上取决于存储实例的数量。 为了改进持续学习的排练性能,特别是少数存储的实例,我们提出了一种新的方法来学习一小套综合范例,捕捉完整数据集的精髓。我们不是直接学习这些合成实例,而是学习每个例子共享组成部分的加权组合,从而大大提高记忆效率。我们展示了我们关于常用数据集的方法的性能,并将其与最近提出的方法和基线进行比较。