Learning from changing tasks and sequential experience without forgetting the obtained knowledge is a challenging problem for artificial neural networks. In this work, we focus on two challenging problems in the paradigm of Continual Learning (CL) without involving any old data: (i) the accumulation of catastrophic forgetting caused by the gradually fading knowledge space from which the model learns the previous knowledge; (ii) the uncontrolled tug-of-war dynamics to balance the stability and plasticity during the learning of new tasks. In order to tackle these problems, we present Progressive Learning without Forgetting (PLwF) and a credit assignment regime in the optimizer. PLwF densely introduces model functions from previous tasks to construct a knowledge space such that it contains the most reliable knowledge on each task and the distribution information of different tasks, while credit assignment controls the tug-of-war dynamics by removing gradient conflict through projection. Extensive ablative experiments demonstrate the effectiveness of PLwF and credit assignment. In comparison with other CL methods, we report notably better results even without relying on any raw data.
翻译:从不断变化的任务和相继经验中学习,而不忘记所获得的知识,对于人工神经网络来说,这是一个具有挑战性的问题。在这项工作中,我们侧重于持续学习范式中的两个具有挑战性的问题,而不涉及任何旧数据:(一) 模型学习以前知识的知识空间逐渐消退,从而造成灾难性的遗忘;(二) 在学习新任务期间,为平衡稳定性和可塑性而无控制的拖动战动态。为了解决这些问题,我们介绍了进步学习而不遗忘(PLwF)和优化者(PLwF)的信用分配制度。PLwF 密集引入了以前任务模式功能,以构建一个知识空间,使其包含关于每项任务的最可靠知识以及不同任务的信息分配,而信用分配则通过预测消除梯度冲突,控制了拖格战争动态。广泛的混合实验表明了PLwF和信用分配的有效性。与其他CL方法相比,我们报告的结果特别好,即使不依赖任何原始数据。