Catastrophic forgetting (CF) happens whenever a neural network overwrites past knowledge while being trained on new tasks. Common techniques to handle CF include regularization of the weights (using, e.g., their importance on past tasks), and rehearsal strategies, where the network is constantly re-trained on past data. Generative models have also been applied for the latter, in order to have endless sources of data. In this paper, we propose a novel method that combines the strengths of regularization and generative-based rehearsal approaches. Our generative model consists of a normalizing flow (NF), a probabilistic and invertible neural network, trained on the internal embeddings of the network. By keeping a single NF throughout the training process, we show that our memory overhead remains constant. In addition, exploiting the invertibility of the NF, we propose a simple approach to regularize the network's embeddings with respect to past tasks. We show that our method performs favorably with espect to state-of-the-art approaches in the literature, with bounded computational power and memory overheads.
翻译:当神经网络在接受新任务培训时,一旦神经网络在以往的知识中重叠,就会发生灾难性的遗忘(CF)。处理CF的常见技术包括重量的正规化(使用过去任务的重要性等)和演练策略,因为网络不断对过去的数据进行再培训。对后者也应用了生成模型,以便获得无穷无尽的数据来源。在本文中,我们提出了一个新颖的方法,将正规化和基于基因的演练方法的优势结合起来。我们的基因模型包括一种正常化流程(NF),一种概率和不可逆的神经网络,在网络的内部嵌入方面受过训练。我们通过在整个培训过程中保持单一的NF,我们表明我们的记忆管理保持不变。此外,我们利用NF的不可忽视性,提出了一种简单的方法来规范网络在以往任务方面的嵌入。我们显示我们的方法与文献中最先进的方法相适应,具有受约束的计算力和记忆管理。