An important problem in machine learning is the ability to learn tasks in a sequential manner. If trained with standard first-order methods most models forget previously learned tasks when trained on a new task, which is often referred to as catastrophic forgetting. A popular approach to overcome forgetting is to regularize the loss function by penalizing models that perform poorly on previous tasks. For example, elastic weight consolidation (EWC) regularizes with a quadratic form involving a diagonal matrix build based on past data. While EWC works very well for some setups, we show that, even under otherwise ideal conditions, it can provably suffer catastrophic forgetting if the diagonal matrix is a poor approximation of the Hessian matrix of previous tasks. We propose a simple approach to overcome this: Regularizing training of a new task with sketches of the Jacobian matrix of past data. This provably enables overcoming catastrophic forgetting for linear models and for wide neural networks, at the cost of memory. The overarching goal of this paper is to provided insights on when regularization-based continual learning algorithms work and under what memory costs.
翻译:机器学习中的一个重要问题是以相继方式学习任务的能力。如果接受标准一阶方法培训后,大多数模型在接受新任务培训时忘记了以前学到的任务,而新任务往往被称为灾难性的遗忘。克服遗忘的流行办法是通过惩罚以往任务执行不良的模式来规范损失功能。例如,弹性重量整合(EWC)以基于过去数据建立对数矩阵的二次形式正规化。尽管EWC对某些设置非常有效,但即使在其他理想条件下,它也可能遭受灾难性的遗忘,如果对正统矩阵与以前任务海珊矩阵的近似不足的话。我们建议了一个简单的方法来克服这一点:用过去数据雅各矩阵的草图对新任务进行正规化培训。这可以证明能够克服线性模型和大神经网络灾难性的遗忘,以记忆为代价。本文件的总体目标是提供关于基于正规化的持续学习算法何时起作用以及记忆成本如何的洞察。