Deep neural networks (DNNs) have recently achieved a great success in computer vision and several related fields. Despite such progress, current neural architectures still suffer from catastrophic interference (a.k.a. forgetting) which obstructs DNNs to learn continually. While several state-of-the-art methods have been proposed to mitigate forgetting, these existing solutions are either highly rigid (as regularization) or time/memory demanding (as replay). An intermediate class of methods, based on dynamic networks, has been proposed in the literature and provides a reasonable balance between task memorization and computational footprint. In this paper, we devise a dynamic network architecture for continual learning based on a novel forgetting-free neural block (FFNB). Training FFNB features on new tasks is achieved using a novel procedure that constrains the underlying parameters in the null-space of the previous tasks, while training classifier parameters equates to Fisher discriminant analysis. The latter provides an effective incremental process which is also optimal from a Bayesian perspective. The trained features and classifiers are further enhanced using an incremental "end-to-end" fine-tuning. Extensive experiments, conducted on different challenging classification problems, show the high effectiveness of the proposed method.
翻译:尽管取得了这些进展,但目前的神经结构仍然受到灾难性干扰(a.k.a.greget),阻碍DNNS不断学习。虽然提出了几种最先进的方法来减轻遗忘,但这些现有解决办法要么非常僵化(正规化),要么要求时间/模棱两可(重现),文献中提出了基于动态网络的中间方法,在任务记忆和计算足迹之间提供了合理的平衡。在本文中,我们设计了一个动态网络结构,以不断学习,这种结构阻碍DNNNP不断学习(FFNB)。关于新任务的培训正在采用新的程序,该程序限制前一项任务的空格空空间的基本参数,而培训分类参数相当于对渔业的干扰性分析。后者提供了一种有效的递增过程,从巴伊西亚的角度来说也是最理想的。经过培训的特征和分类人员正在使用一种渐进式的“端到端”的微调方法来进一步强化。在高层次上进行不同的分类,展示了拟议的高难度性实验。