Catastrophic forgetting has been a significant problem hindering the deployment of deep learning algorithms in the continual learning setting. Numerous methods have been proposed to address the catastrophic forgetting problem where an agent loses its generalization power of old tasks while learning new tasks. We put forward an alternative strategy to handle the catastrophic forgetting with knowledge amalgamation (CFA), which learns a student network from multiple heterogeneous teacher models specializing in previous tasks and can be applied to current offline methods. The knowledge amalgamation process is carried out in a single-head manner with only a selected number of memorized samples and no annotations. The teachers and students do not need to share the same network structure, allowing heterogeneous tasks to be adapted to a compact or sparse data representation. We compare our method with competitive baselines from different strategies, demonstrating our approach's advantages.
翻译:灾难性的遗忘是阻碍在不断学习环境中运用深层次学习算法的一个严重问题,已经提出了许多方法来解决灾难性的遗忘问题,即代理人在学习新任务的同时丧失了对旧任务的普遍化能力。我们提出了一个替代战略,用知识结合(CFA)来处理灾难性的遗忘问题。CFA从多种不同教师模式中学习了专门从事以往任务的多式教师网络,并可用于当前的离线方法。知识合并进程以单头方式进行,只选用若干记忆样本和无说明。教师和学生不需要共享相同的网络结构,让不同的任务适应紧凑或稀少的数据代表。我们比较了我们的方法与不同战略的竞争基线,显示了我们的方法的优势。