The ability to sequentially learn multiple tasks without forgetting is a key skill of biological brains, whereas it represents a major challenge to the field of deep learning. To avoid catastrophic forgetting, various continual learning (CL) approaches have been devised. However, these usually require discrete task boundaries. This requirement seems biologically implausible and often limits the application of CL methods in the real world where tasks are not always well defined. Here, we take inspiration from neuroscience, where sparse, non-overlapping neuronal representations have been suggested to prevent catastrophic forgetting. As in the brain, we argue that these sparse representations should be chosen on the basis of feed forward (stimulus-specific) as well as top-down (context-specific) information. To implement such selective sparsity, we use a bio-plausible form of hierarchical credit assignment known as Deep Feedback Control (DFC) and combine it with a winner-take-all sparsity mechanism. In addition to sparsity, we introduce lateral recurrent connections within each layer to further protect previously learned representations. We evaluate the new sparse-recurrent version of DFC on the split-MNIST computer vision benchmark and show that only the combination of sparsity and intra-layer recurrent connections improves CL performance with respect to standard backpropagation. Our method achieves similar performance to well-known CL methods, such as Elastic Weight Consolidation and Synaptic Intelligence, without requiring information about task boundaries. Overall, we showcase the idea of adopting computational principles from the brain to derive new, task-free learning algorithms for CL.
翻译:连续学习多种任务而不忘记的能力是生物大脑的关键技能,而这是深层学习领域的一大挑战。为了避免灾难性的遗忘,已经设计了各种持续学习(CL)方法。然而,这些要求通常需要不同的任务界限。这种要求似乎在生物学上不可信,而且往往限制在现实世界中应用CL方法,因为在现实世界中,任务往往没有很好地界定。这里,我们从神经科学中汲取灵感,在神经科学中建议了稀少的、不重叠的神经表现方式,以防止灾难性的遗忘。就像在大脑中一样,我们主张这些稀少的表述方式应该根据反馈(特定模拟)和自上而下(特定文本)信息来选择。为了实施这种选择性的宽度,我们使用一种生物可塑性的信用分配形式,称为深度反馈控制(DFC),并将它与赢取的全局性松动机制结合起来。除了紧张外,我们还在每一层中引入了更熟悉的重复的连接,以进一步保护先前的表达方式。我们评估了不同-MICTFC-C-C-C-C-Slical recal recal recal recal recal recal recal real real requistrismactalation,我们只能以达到相同的C-hal 和Sy 和C-hal-hal-hal-hal-hal-hutdal-hutdal-hlutdal-s-s to to 和我们只使用一种我们使用标准的周期性方法, 。