Animals thrive in a constantly changing environment and leverage the temporal structure to learn well-factorized causal representations. In contrast, traditional neural networks suffer from forgetting in changing environments and many methods have been proposed to limit forgetting with different trade-offs. Inspired by the brain thalamocortical circuit, we introduce a simple algorithm that uses optimization at inference time to generate internal representations of temporal context and to infer current context dynamically, allowing the agent to parse the stream of temporal experience into discrete events and organize learning about them. We show that a network trained on a series of tasks using traditional weight updates can infer tasks dynamically using gradient descent steps in the latent task embedding space (latent updates). We then alternate between the weight updates and the latent updates to arrive at Thalamus, a task-agnostic algorithm capable of discovering disentangled representations in a stream of unlabeled tasks using simple gradient descent. On a continual learning benchmark, it achieves competitive end average accuracy and demonstrates knowledge transfer. After learning a subset of tasks it can generalize to unseen tasks as they become reachable within the well-factorized latent space, through one-shot latent updates. The algorithm meets many of the desiderata of an ideal continually learning agent in open-ended environments, and its simplicity suggests fundamental computations in circuits with abundant feedback control loops such as the thalamocortical circuits in the brain.
翻译:动物在不断变化的环境中繁衍,利用时间结构来学习各种因素造成的因果关系。相比之下,传统神经网络在不断变化的环境中被遗忘,而传统神经网络则在不断变化的环境中被遗忘,并提出了许多方法来限制以不同的权衡方式忘记。在大脑 Thalamocortical 电路的启发下,我们引入了一个简单的算法,在推论时间里使用最优化来产生时间背景的内部表述,并动态地推断当前环境,使代理商能够将时间经验流分析成离散事件,并组织知识转移。我们表明,在使用传统重量更新进行一系列培训的网络中,可以动态地推移任务,在潜在任务嵌入空间(延后更新)中,使用梯度递增步骤来限制对时间的遗忘。我们随后在重量更新和潜在更新之间进行交替,以便到达Thalamamus, 一种任务 — 任务- 任务- 任务性算算算法能够发现无标签的任务流中分解的表达方式, 通过持续学习基准, 达到具有竞争性的终端平均准确性, 并显示知识转移。我们学习了一组任务可以概括到在深层层层空间中可以达到可达目的的深层回路路流, 。