Animals thrive in a constantly changing environment and leverage the temporal structure to learn well-factorized causal representations. In contrast, traditional neural networks suffer from forgetting in changing environments and many methods have been proposed to limit forgetting with different trade-offs. Inspired by the brain thalamocortical circuit, we introduce a simple algorithm that uses optimization at inference time to generate internal representations of the current task dynamically. The algorithm alternates between updating the model weights and a latent task embedding, allowing the agent to parse the stream of temporal experience into discrete events and organize learning about them. On a continual learning benchmark, it achieves competitive end average accuracy by mitigating forgetting, but importantly, by requiring the model to adapt through latent updates, it organizes knowledge into flexible structures with a cognitive interface to control them. Tasks later in the sequence can be solved through knowledge transfer as they become reachable within the well-factorized latent space. The algorithm meets many of the desiderata of an ideal continually learning agent in open-ended environments, and its simplicity suggests fundamental computations in circuits with abundant feedback control loops such as the thalamocortical circuits in the brain.
翻译:动物在不断变化的环境中生长,并利用时间结构来学习有良好因素的因果关系。相反,传统神经网络在变化的环境中被遗忘,在不断变化的环境中被遗忘,并且提出了许多方法来限制以不同权衡方式的遗忘。在大脑Thalamocortal电路的启发下,我们引入了一种简单的算法,在推论时间使用优化来产生对当前任务动态的内部表述。算法在更新模型重量和潜在嵌入任务之间互为替代,使代理人能够将时间经验流分解为离散事件,并组织关于这些事件的学习。在不断学习的基准中,它通过减少遗忘而实现具有竞争力的终端平均准确性,但重要的是,它通过要求模型通过潜伏更新来适应,将知识组织成灵活的结构,并有一个认知界面来控制这些变化。在后期的任务可以通过知识转移来解答,因为它们在有良好因素的潜伏潜伏层空间中可以接触到。这一算法符合一个理想的持续学习剂在开放环境中的脱缘关系,其简单性表明在使用大量反馈控制循环中进行基本计算。</s>