In lifelong learning systems based on artificial neural networks, one of the biggest obstacles is the inability to retain old knowledge as new information is encountered. This phenomenon is known as catastrophic forgetting. In this paper, we propose a new kind of connectionist architecture, the Sequential Neural Coding Network, that is robust to forgetting when learning from streams of data points and, unlike networks of today, does not learn via the popular back-propagation of errors. Grounded in the neurocognitive theory of predictive processing, our model adapts synapses in a biologically-plausible fashion while another neural system learns to direct and control this cortex-like structure, mimicking some of the task-executive control functionality of the basal ganglia. In our experiments, we demonstrate that our self-organizing system experiences significantly less forgetting compared to standard neural models, outperforming a swath of previously proposed methods, including rehearsal/data buffer-based methods, on both standard (SplitMNIST, Split Fashion MNIST, etc.) and custom benchmarks even though it is trained in a stream-like fashion. Our work offers evidence that emulating mechanisms in real neuronal systems, e.g., local learning, lateral competition, can yield new directions and possibilities for tackling the grand challenge of lifelong machine learning.
翻译:在以人工神经网络为基础的终身学习系统中,最大的障碍之一是无法保留旧知识,因为新信息会遇到。这一现象被称为灾难性的遗忘。在本文中,我们提出一种新的连接型结构,即序列神经编码网络,在从数据流中学习时能够忘却,与今天的网络不同,它不会通过流行的对错误的反反向分析学习。基于预测处理的神经认知理论,我们的模型以生物可视的方式调整神经突触,而另一个神经系统则学习指导和控制这种像皮层的结构,模仿核磁共振的一些任务-执行控制功能。在我们的实验中,我们证明我们的自我组织系统与标准的神经模型相比,远远没有忘记,比以往提出的方法(包括演练/数据缓冲方法)的一丝不苟。我们的工作可以提供大量经验,既基于标准(SplidMIT,Slip Fashon MNIST,等等),又基于定制基准,即使它以流式的方式培训,模拟着一些任务-执行控制结构的功能。我们在实验中,我们的工作可以提供新的证据,用来解决巨型的系统。