While biological intelligence grows organically as new knowledge is gathered throughout life, Artificial Neural Networks forget catastrophically whenever they face a changing training data distribution. Rehearsal-based Continual Learning (CL) approaches have been established as a versatile and reliable solution to overcome this limitation; however, sudden input disruptions and memory constraints are known to alter the consistency of their predictions. We study this phenomenon by investigating the geometric characteristics of the learner's latent space and find that replayed data points of different classes increasingly mix up, interfering with classification. Hence, we propose a geometric regularizer that enforces weak requirements on the Laplacian spectrum of the latent space, promoting a partitioning behavior. We show that our proposal, called Continual Spectral Regularizer (CaSpeR), can be easily combined with any rehearsal-based CL approach and improves the performance of SOTA methods on standard benchmarks. Finally, we conduct additional analysis to provide insights into CaSpeR's effects and applicability.
翻译:尽管生物智能随着新知识在整个生命中不断积累而有机地增长,人工神经网络在面临不断变化的培训数据分布时会忘却灾难性地忘记了每时每刻都在变化的训练数据分布。以排练为基础的连续学习(CL)方法已被确定为克服这一限制的多种和可靠的解决办法;然而,人们知道,突然输入中断和记忆限制会改变其预测的一致性。我们通过调查学习者潜在空间的几何特征来研究这一现象,发现重播不同类别的数据点越来越混杂,从而干扰分类。因此,我们提议了一个几何定调器,对潜空的拉普拉奇频谱执行薄弱的要求,促进分层行为。我们显示,我们的提案,即所谓的“CaSper”(CaSpeR),很容易与任何以排练为基础的CL(CL)方法相结合,并改进SOTA方法在标准基准方面的性能。最后,我们进行更多的分析,以提供对CASpeR的影响和适用性。