We propose a novel regularization-based continual learning method, dubbed as Adaptive Group Sparsity based Continual Learning (AGS-CL), using two group sparsity-based penalties. Our method selectively employs the two penalties when learning each node based its the importance, which is adaptively updated after learning each new task. By utilizing the proximal gradient descent method for learning, the exact sparsity and freezing of the model is guaranteed, and thus, the learner can explicitly control the model capacity as the learning continues. Furthermore, as a critical detail, we re-initialize the weights associated with unimportant nodes after learning each task in order to prevent the negative transfer that causes the catastrophic forgetting and facilitate efficient learning of new tasks. Throughout the extensive experimental results, we show that our AGS-CL uses much less additional memory space for storing the regularization parameters, and it significantly outperforms several state-of-the-art baselines on representative continual learning benchmarks for both supervised and reinforcement learning tasks.
翻译:我们提出一种新的基于正规化的持续学习方法,称为基于适应性群体分化的连续学习(AGS-CL),使用两个基于群体宽度的处罚。我们的方法在学习每个节点时有选择地使用两种惩罚,以其重要性为基础,在学习每一项新任务后加以适应性更新。通过使用最接近的梯度梯度下降方法进行学习,模型的准确宽度和冻结得到保证,因此,学习者可以随着学习的继续明确控制模型能力。此外,作为一个关键的细节,我们重新启用与每个任务学习之后不重要的节点相关的重量,以防止造成灾难性遗忘的负转移,并促进高效地学习新任务。我们在整个广泛的实验结果中,AGS-CL使用更多的记忆空间来存储规范化参数,大大超过监督性和强化学习任务的有代表性的持续学习基准。