Meta-learning algorithms leverage regularities that are present on a set of tasks to speed up and improve the performance of a subsidiary learning process. Recent work on deep neural networks has shown that prior gradient-based learning of meta-parameters can greatly improve the efficiency of subsequent learning. Here, we present a gradient-based meta-learning algorithm based on equilibrium propagation. Instead of explicitly differentiating the learning process, our contrastive meta-learning rule estimates meta-parameter gradients by executing the subsidiary process more than once. This avoids reversing the learning dynamics in time and computing second-order derivatives. In spite of this, and unlike previous first-order methods, our rule recovers an arbitrarily accurate meta-parameter update given enough compute. As such, contrastive meta-learning is a candidate rule for biologically-plausible meta-learning. We establish theoretical bounds on its performance and present experiments on a set of standard benchmarks and neural network architectures.
翻译:元学习算法利用在一系列任务上存在的规律来加快和改进辅助学习过程的绩效。关于深神经网络的近期工作表明,以前基于梯度的元参数学习可以大大提高后续学习的效率。在这里,我们提出了一个基于均衡传播的基于梯度的元学习算法。我们没有明确区分学习过程,相反的元学习规则通过不止一次地执行辅助过程来估计元参数梯。这避免了在时间和计算二阶衍生物方面的学习动态发生逆转。尽管如此,而且与先前的第一阶方法不同,我们的规则恢复了一种任意准确的元参数更新,并有足够的精度。由于这种对比式的元学习是生物可塑性元学习的候选规则。我们建立了有关其业绩的理论界限,并提出了一套标准基准和神经网络结构的实验。