The mainstream paradigm behind continual learning has been to adapt the model parameters to non-stationary data distributions, where catastrophic forgetting is the central challenge. Typical methods rely on a rehearsal buffer or known task identity at test time to retrieve learned knowledge and address forgetting, while this work presents a new paradigm for continual learning that aims to train a more succinct memory system without accessing task identity at test time. Our method learns to dynamically prompt (L2P) a pre-trained model to learn tasks sequentially under different task transitions. In our proposed framework, prompts are small learnable parameters, which are maintained in a memory space. The objective is to optimize prompts to instruct the model prediction and explicitly manage task-invariant and task-specific knowledge while maintaining model plasticity. We conduct comprehensive experiments under popular image classification benchmarks with different challenging continual learning settings, where L2P consistently outperforms prior state-of-the-art methods. Surprisingly, L2P achieves competitive results against rehearsal-based methods even without a rehearsal buffer and is directly applicable to challenging task-agnostic continual learning. Source code is available at https://github.com/google-research/l2p.
翻译:持续学习背后的主流模式是将模型参数调整到非静止数据分布中,灾难性的遗忘是核心挑战。典型的方法依靠测试时的预演缓冲或已知任务身份来获取学到的知识并解决遗忘问题,而这项工作则为持续学习提供了一个新的范例,目的是在测试时训练一个更加简洁的记忆系统,而没有获得任务身份。我们的方法是动态地迅速(L2P)一个经过预先培训的模型,以便在不同任务过渡期间按顺序学习任务。在我们提议的框架中,提示是小的可学习参数,在记忆空间中保留。目标是优化提示,以指导模型预测,明确管理任务差异性和任务特有知识,同时保持模型的可塑性。我们在流行图像分类基准下进行全面实验,有不同的持续学习环境,即L2P持续地超越先前的状态方法。令人惊讶的是,L2P在没有彩排缓冲的情况下,在彩排方法下取得竞争性的结果,直接适用于具有挑战性的任务持续学习。源代码可在 https://githubus.com/golears-searching-sing.