Our brains extract durable, generalizable knowledge from transient experiences of the world. Artificial neural networks come nowhere close: when tasked with learning to classify objects by training on non-repeating video frames in temporal order (online stream learning), models that learn well from shuffled datasets catastrophically forget old knowledge upon learning new stimuli. We propose a new continual learning algorithm, Compositional Replay Using Memory Blocks (CRUMB), which mitigates forgetting by replaying feature maps reconstructed by recombining generic parts. Just as crumbs together form a loaf of bread, we concatenate trainable and re-usable "memory block" vectors to compositionally reconstruct feature map tensors in convolutional neural networks. CRUMB stores the indices of memory blocks used to reconstruct new stimuli, enabling replay of specific memories during later tasks. CRUMB's memory blocks are tuned to enhance replay: a single feature map stored, reconstructed, and replayed by CRUMB mitigates forgetting during video stream learning more effectively than an entire image, even though it occupies only 3.6% as much memory. We stress-tested CRUMB alongside 13 competing methods on 5 challenging datasets. To address the limited number of existing online stream learning datasets, we introduce 2 new benchmarks by adapting existing datasets for stream learning. With about 4% of the memory and 20% of the runtime, CRUMB mitigates catastrophic forgetting more effectively than the prior state-of-the-art. Our code is available at https://github.com/MorganBDT/crumb.git.
翻译:我们的大脑从世界的短暂经历中提取了耐久的、可概括的知识。人工神经网络远没有那么接近:当我们的任务是通过在不复制的视频框架上进行时间顺序(在线流学习)的培训来学习对对象进行分类时,从打乱的数据集中学习的模型灾难性地忘记了旧的知识。我们建议了一种新的持续学习算法,即“利用记忆区块(CRUMB)”,它通过重新播放通过重新组合通用部件重建的功能图来减轻遗忘。正像碎块一起形成一个面包块一样,我们将可培训和可重新使用的“模版块”矢量矢量进行分类,以组成重建革命性神经网络中的地貌图。CRUMB 储存用于重建新模量的记忆区块指数,在后期任务中可以重新显示具体的记忆区块。CRUB的单一特性图被存储、重建以及由CRUMB重新显示的状态,在磁盘流学习过程中比整个图像更有效,尽管它只是不断缩小了CMB的数据流中的3.6%。我们目前的数据压力学习了目前的数据。</s>