Artificial neural networks (ANNs) are typically confined to accomplishing pre-defined tasks by learning a set of static parameters. In contrast, biological neural networks (BNNs) can adapt to various new tasks by continually updating the neural connections based on the inputs, which is aligned with the paradigm of learning effective learning rules in addition to static parameters, e.g., meta-learning. Among various biologically inspired learning rules, Hebbian plasticity updates the neural network weights using local signals without the guide of an explicit target function, thus enabling an agent to learn automatically without human efforts. However, typical plastic ANNs using a large amount of meta-parameters violate the nature of the genomics bottleneck and potentially deteriorate the generalization capacity. This work proposes a new learning paradigm decomposing those connection-dependent plasticity rules into neuron-dependent rules thus accommodating $\Theta(n^2)$ learnable parameters with only $\Theta(n)$ meta-parameters. We also thoroughly study the effect of different neural modulation on plasticity. Our algorithms are tested in challenging random 2D maze environments, where the agents have to use their past experiences to shape the neural connections and improve their performances for the future. The results of our experiment validate the following: 1. Plasticity can be adopted to continually update a randomly initialized RNN to surpass pre-trained, more sophisticated recurrent models, especially when coming to long-term memorization. 2. Following the genomics bottleneck, the proposed decomposed plasticity can be comparable to or even more effective than canonical plasticity rules in some instances.
翻译:人工神经网络(ANNS)通常局限于通过学习一组静态参数来完成预先确定的任务。相反,生物神经网络(BNNS)可以通过不断更新基于投入的神经连接来适应各种新任务,这些输入与学习有效学习规则的范式相一致,并且与学习静态参数(例如元学习)相匹配。在各种生物启发的学习规则中,Hebbian可塑性利用当地信号更新神经网络重量,而没有明确的目标功能指南,从而使一个代理可以自动学习,而无需人类的努力。然而,生物神经网络(BNNS)可以不断更新基于投入的神经连接,这符合学习有效学习规则的范式。在适应 $Theta(n) 的情况下,Hebbian塑料网络的可学习参数只能使用$Theta(n) 元直径直径的参数,这样可以让一个复合模型自动学习,这样可以让一个代理人自动学习。但是,即使使用大量的元参数,也能够彻底研究不同的神经模型对可自动学习的影响。 我们的定序后,特别是变式的变式规则可以持续地测试到过去。