Equilibrium systems are a powerful way to express neural computations. As special cases, they include models of great current interest in both neuroscience and machine learning, such as deep neural networks, equilibrium recurrent neural networks, deep equilibrium models, or meta-learning. Here, we present a new principle for learning such systems with a temporally- and spatially-local rule. Our principle casts learning as a least-control problem, where we first introduce an optimal controller to lead the system towards a solution state, and then define learning as reducing the amount of control needed to reach such a state. We show that incorporating learning signals within a dynamics as an optimal control enables transmitting activity-dependent credit assignment information, avoids storing intermediate states in memory, and does not rely on infinitesimal learning signals. In practice, our principle leads to strong performance matching that of leading gradient-based learning methods when applied to an array of problems involving recurrent neural networks and meta-learning. Our results shed light on how the brain might learn and offer new ways of approaching a broad class of machine learning problems.
翻译:平衡系统是表达神经计算的一种强有力的方法。 作为特殊案例,它们包括当前对神经科学和机器学习都非常感兴趣的模型,如深神经网络、平衡的经常性神经网络、深平衡模型或元学习。在这里,我们提出了一个学习这种系统的新原则,它具有时间和空间上的规律。我们的原则将学习作为一个最不控制的问题,我们首先引入一个最佳控制器,引导系统走向一个解决方案状态,然后将学习定义为减少达到这样的状态所需的控制量。我们表明,将学习信号纳入动态中作为最佳控制,能够传递依赖活动的信用分配信息,避免存储记忆中的中间状态,不依赖无限的学习信号。在实践中,我们的原则导致在应用到一系列涉及经常性神经网络和元学习的问题时,与基于梯度的学习方法高度匹配。我们的结果揭示了大脑如何学习和提供新的方法,以接近广泛的机器学习问题。