Continual learning of a stream of tasks is an active area in deep neural networks. The main challenge investigated has been the phenomenon of catastrophic forgetting or interference of newly acquired knowledge with knowledge from previous tasks. Recent work has investigated forward knowledge transfer to new tasks. Backward transfer for improving knowledge gained during previous tasks has received much less attention. There is in general limited understanding of how knowledge transfer could aid tasks learned continually. We present a theory for knowledge transfer in continual supervised learning, which considers both forward and backward transfer. We aim at understanding their impact for increasingly knowledgeable learners. We derive error bounds for each of these transfer mechanisms. These bounds are agnostic to specific implementations (e.g. deep neural networks). We demonstrate that, for a continual learner that observes related tasks, both forward and backward transfer can contribute to an increasing performance as more tasks are observed.
翻译:不断学习一串任务是深层神经网络的一个活跃领域,所调查的主要挑战是灾难性地忘记或干扰以以前任务的知识获得的新知识的现象;最近的工作调查了向新任务转移知识的先期研究;前期工作所获知识的后向转移得到的后向转移得到的注意少得多;一般而言,对知识转让如何有助于不断学习任务的理解有限;我们提出在不断监督的学习中进行知识转让的理论,既考虑前向转让,又考虑后向转移;我们的目的是了解这些知识日益丰富的学习者的影响;我们为这些转让机制的每一个机制找出错误界限。这些界限对具体执行(例如深神经网络)是不可知的。我们证明,对于不断学习、观察相关任务的学习者来说,前向和后向转移都有助于不断提高业绩,因为看到更多的任务。