Biological agents are known to learn many different tasks over the course of their lives, and to be able to revisit previous tasks and behaviors with little to no loss in performance. In contrast, artificial agents are prone to 'catastrophic forgetting' whereby performance on previous tasks deteriorates rapidly as new ones are acquired. This shortcoming has recently been addressed using methods that encourage parameters to stay close to those used for previous tasks. This can be done by (i) using specific parameter regularizers that map out suitable destinations in parameter space, or (ii) guiding the optimization journey by projecting gradients into subspaces that do not interfere with previous tasks. However, parameter regularization has been shown to be relatively ineffective in recurrent neural networks (RNNs), a setting relevant to the study of neural dynamics supporting biological continual learning. Similarly, projection based methods can reach capacity and fail to learn any further as the number of tasks increases. To address these limitations, we propose Natural Continual Learning (NCL), a new method that unifies weight regularization and projected gradient descent. NCL uses Bayesian weight regularization to encourage good performance on all tasks at convergence and combines this with gradient projections designed to prevent catastrophic forgetting during optimization. NCL formalizes gradient projection as a trust region algorithm based on the Fisher information metric, and achieves scalability via a novel Kronecker-factored approximation strategy. Our method outperforms both standard weight regularization techniques and projection based approaches when applied to continual learning problems in RNNs. The trained networks evolve task-specific dynamics that are strongly preserved as new tasks are learned, similar to experimental findings in biological circuits.
翻译:已知生物物剂在生命过程中会学习许多不同的任务,并能重新审视以往的任务和行为,而其性能很少甚至没有损失。相反,人造物剂容易被“灾难性的忘记”而“灾难性的忘记”导致随着新任务获得,先前任务的绩效会随着新任务而迅速恶化。最近,利用鼓励参数与前任务所用参数保持接近的方法解决了这一缺陷。为了消除这些局限性,我们建议采用具体参数规范,在参数空间中绘制适当的目的地,或者(二)通过预测梯度进入不干扰先前任务的子空间来指导优化旅程。然而,在经常性神经网络(RNNN)中,参数的正规化被证明相对无效,而对于支持生物持续学习的神经动态的研究也与此相关。同样,基于预测的方法可以达到能力,但随着任务数量的增加,我们建议采用自然持续学习(NCL)这一新的方法来统一重量正规化和预测梯度。 NCL 使用比重来鼓励所有任务的正常化,在不断的神经力网络中,将经过训练的精确性变压方法结合起来,从而在不断变压的不断变压中,从而防止以不断变压的方法在不断变压中实现。