Equilibrium systems are a powerful way to express neural computations. As special cases, they include models of great current interest in both neuroscience and machine learning, such as equilibrium recurrent neural networks, deep equilibrium models, or meta-learning. Here, we present a new principle for learning such systems with a temporally- and spatially-local rule. Our principle casts learning as a least-control problem, where we first introduce an optimal controller to lead the system towards a solution state, and then define learning as reducing the amount of control needed to reach such a state. We show that incorporating learning signals within a dynamics as an optimal control enables transmitting credit assignment information in previously unknown ways, avoids storing intermediate states in memory, and does not rely on infinitesimal learning signals. In practice, our principle leads to strong performance matching that of leading gradient-based learning methods when applied to an array of problems involving recurrent neural networks and meta-learning. Our results shed light on how the brain might learn and offer new ways of approaching a broad class of machine learning problems.
翻译:平衡系统是表达神经计算的一种强有力的方法。 作为特殊案例,它们包括当前对神经科学和机器学习都非常感兴趣的模型,例如平衡的经常性神经网络、深平衡模型或元学习。在这里,我们提出了一个学习这种系统的新原则,它具有时间和空间上的当地规则。我们的原则将学习作为一个控制最弱的问题,我们首先引入一个最佳控制器,引导系统走向一个解决方案状态,然后将学习定义为减少达到这样的状态所需的控制量。我们表明,将学习信号纳入动态中,作为最佳控制,能够以以前未知的方式传输信用分配信息,避免将中间状态存储在记忆中,不依赖无限的学习信号。在实践中,我们的原则导致在应用到一系列涉及经常性神经网络和元学习的问题时,与引导基于梯度的学习方法相匹配。我们的结果揭示了大脑如何学习和提供新的方法来接近广泛的机器学习问题。