Recent self-supervised learning methods are able to learn high-quality image representations and are closing the gap with supervised methods. However, these methods are unable to acquire new knowledge incrementally -- they are, in fact, mostly used only as a pre-training phase with IID data. In this work we investigate self-supervised methods in continual learning regimes without additional memory or replay. To prevent forgetting of previous knowledge, we propose the usage of functional regularization. We will show that naive functional regularization, also known as feature distillation, leads to low plasticity and therefore seriously limits continual learning performance. To address this problem, we propose Projected Functional Regularization where a separate projection network ensures that the newly learned feature space preserves information of the previous feature space, while allowing for the learning of new features. This allows us to prevent forgetting while maintaining the plasticity of the learner. Evaluation against other incremental learning approaches applied to self-supervision demonstrates that our method obtains competitive performance in different scenarios and on multiple datasets.
翻译:最近自我监督的学习方法能够学习高质量的图像表现,并正在通过监督方法缩小差距。然而,这些方法无法逐步获得新的知识 -- -- 事实上,这些方法大多仅用作使用ID数据的培训前阶段。在这项工作中,我们调查了连续学习制度中自我监督的方法,而没有额外的记忆或重播。为防止忘记先前的知识,我们提议使用功能正规化。我们将显示天真的功能正规化,又称为特征蒸馏,导致低塑料性,从而严重限制持续学习的绩效。为了解决这个问题,我们建议了预测功能正规化,因为一个单独的预测网络确保新学到的功能空间保留了先前特征空间的信息,同时允许学习新的特征。这使我们能够在保持学习者的塑料特性的同时防止忘记。对照适用于自我监督的其他渐进学习方法进行的评估表明,我们的方法在不同情景和多个数据集中取得了竞争性的性表现。