Meta-learning algorithms leverage regularities that are present on a set of tasks to speed up and improve the performance of a subsidiary learning process. Recent work on deep neural networks has shown that prior gradient-based learning of meta-parameters can greatly improve the efficiency of subsequent learning. Here, we present a biologically plausible meta-learning algorithm based on equilibrium propagation. Instead of explicitly differentiating the learning process, our contrastive meta-learning rule estimates meta-parameter gradients by executing the subsidiary process more than once. This avoids reversing the learning dynamics in time and computing second-order derivatives. In spite of this, and unlike previous first-order methods, our rule recovers an arbitrarily accurate meta-parameter update given enough compute. We establish theoretical bounds on its performance and present experiments on a set of standard benchmarks and neural network architectures.
翻译:元学习算法利用在一系列任务上存在的规律来加快和改进辅助学习过程的绩效。关于深神经网络的近期工作表明,以前基于梯度的元参数学习可以大大提高后续学习的效率。在这里,我们提出了一个基于均衡传播的生物上可信的元学习算法。我们没有明确区分学习过程,相反的元学习规则通过不止一次执行辅助过程来估计元参数梯度。这避免了在时间和计算二阶衍生物方面的学习动态发生逆转。尽管如此,我们的规则与先前的第一阶方法不同,但又恢复了任意准确的元参数更新,并有足够的精度。我们建立了有关其业绩的理论界限,并提出了一套标准基准和神经网络结构的实验。