This paper proposes an assessor-guided learning strategy for continual learning where an assessor guides the learning process of a base learner by controlling the direction and pace of the learning process thus allowing an efficient learning of new environments while protecting against the catastrophic interference problem. The assessor is trained in a meta-learning manner with a meta-objective to boost the learning process of the base learner. It performs a soft-weighting mechanism of every sample accepting positive samples while rejecting negative samples. The training objective of a base learner is to minimize a meta-weighted combination of the cross entropy loss function, the dark experience replay (DER) loss function and the knowledge distillation loss function whose interactions are controlled in such a way to attain an improved performance. A compensated over-sampling (COS) strategy is developed to overcome the class imbalanced problem of the episodic memory due to limited memory budgets. Our approach, Assessor-Guided Learning Approach (AGLA), has been evaluated in the class-incremental and task-incremental learning problems. AGLA achieves improved performances compared to its competitors while the theoretical analysis of the COS strategy is offered. Source codes of AGLA, baseline algorithms and experimental logs are shared publicly in \url{https://github.com/anwarmaxsum/AGLA} for further study.
翻译:本文提出了一种面向连续学习的评估者引导学习策略,其中评估者通过控制学习过程的方向和速度来引导基础学习者的学习过程,从而实现对新环境的有效学习,并防止灾难性干扰问题的发生。评估者采用元学习方法进行训练,其元目标是增强基础学习者的学习过程。它执行每个样本的软加权机制,接受正样本并拒绝负样本。基础学习者的训练目标是最小化交叉熵损失函数、暗记忆重放(DER)损失函数和知识蒸馏损失函数的元加权组合,这些交互以改善性能为目的进行控制。开发了一种补偿过采样(COS)策略来克服由于有限内存预算而导致的叙事性记忆的类不平衡问题。我们的方法Assessor-Guided Learning Approach(AGLA)在类增量和任务增量学习问题上进行了评估。与其竞争对手相比,AGLA取得了更好的性能,同时提供COS策略的理论分析。AGLA的源代码、基准算法和实验日志在\url{https://github.com/anwarmaxsum/AGLA}上公开分享以供进一步研究。