Modern representation learning methods may fail to adapt quickly under non-stationarity since they suffer from the problem of catastrophic forgetting and decaying plasticity. Such problems prevent learners from fast adaptation to changes since they result in increasing numbers of saturated features and forgetting useful features when presented with new experiences. Hence, these methods are rendered ineffective for continual learning. This paper proposes Utility-based Perturbed Gradient Descent (UPGD), an online representation-learning algorithm well-suited for continual learning agents with no knowledge about task boundaries. UPGD protects useful weights or features from forgetting and perturbs less useful ones based on their utilities. Our empirical results show that UPGD alleviates catastrophic forgetting and decaying plasticity, enabling modern representation learning methods to work in the continual learning setting.
翻译:现代代议制学习方法可能无法在非常态情况下迅速适应,因为他们面临灾难性的遗忘和腐烂的塑料问题,这些问题使学习者无法迅速适应变化,因为这些问题造成越来越多的饱和特征,在介绍新经验时忘记了有用的特征。因此,这些方法对继续学习没有效果。本文件提出,以实用为基础的有固定作用的梯子后裔(UPGD)是一种在线代议制学习算法,适合对任务界限没有了解的不断学习的代理商。UPGD保护有用的重量或特征,使其不因公用设施而忘却和侵扰作用较小的重量或特征。我们的经验结果表明,UPGD减轻了灾难性的遗忘和腐烂的塑料性,使现代代议制学习方法能够在持续学习环境中发挥作用。