Humans and other animals are capable of improving their learning performance as they solve related tasks from a given problem domain, to the point of being able to learn from extremely limited data. While synaptic plasticity is generically thought to underlie learning in the brain, the precise neural and synaptic mechanisms by which learning processes improve through experience are not well understood. Here, we present a general-purpose, biologically-plausible meta-learning rule which estimates gradients with respect to the parameters of an underlying learning algorithm by simply running it twice. Our rule may be understood as a generalization of contrastive Hebbian learning to meta-learning and notably, it neither requires computing second derivatives nor going backwards in time, two characteristic features of previous gradient-based methods that are hard to conceive in physical neural circuits. We demonstrate the generality of our rule by applying it to two distinct models: a complex synapse with internal states which consolidate task-shared information, and a dual-system architecture in which a primary network is rapidly modulated by another one to learn the specifics of each task. For both models, our meta-learning rule matches or outperforms reference algorithms on a wide range of benchmark problems, while only using information presumed to be locally available at neurons and synapses. We corroborate these findings with a theoretical analysis of the gradient estimation error incurred by our rule.
翻译:人类和其他动物在从某个问题领域解决相关任务时,能够提高学习成绩,从某个问题领域解决相关任务,到能够从极其有限的数据学习。虽然合成塑料一般被认为是大脑学习的基础,但人们并不十分理解通过经验改进学习过程的精确神经和合成机制。这里,我们提出了一个通用的、生物可复制的元学习规则,该规则仅用两次运行来估计基本学习算法参数的梯度。我们的规则可以被理解为与Hebbbian相对应的学习向元化学习的概括化,特别是,它既不需要计算第二个衍生物,也不需要向后退,而通常认为大脑的学习基础是精确的神经和合成机制。我们通过将规则应用于两种不同的模型来显示我们规则的普遍性:一个复杂的、生物可复制的元化的元化元学习规则,以及一个由另一个系统结构迅速调整以学习每项任务的具体内容。对于两种模型来说,我们的元化规则都不需要计算第二个衍生物,或者在时间上向后退去。我们以往的梯度方法的两个特征特征特征特征特征特征特征特征,在物理神经变的模型分析中,我们只能使用一个地方的模型的模型的推算算。