Continual learning aims to learn on non-stationary data streams without catastrophically forgetting previous knowledge. Prevalent replay-based methods address this challenge by rehearsing on a small buffer holding the seen data, for which a delicate sample selection strategy is required. However, existing selection schemes typically seek only to maximize the utility of the ongoing selection, overlooking the interference between successive rounds of selection. Motivated by this, we dissect the interaction of sequential selection steps within a framework built on influence functions. We manage to identify a new class of second-order influences that will gradually amplify incidental bias in the replay buffer and compromise the selection process. To regularize the second-order effects, a novel selection objective is proposed, which also has clear connections to two widely adopted criteria. Furthermore, we present an efficient implementation for optimizing the proposed criterion. Experiments on multiple continual learning benchmarks demonstrate the advantage of our approach over state-of-the-art methods. Code is available at https://github.com/feifeiobama/InfluenceCL.
翻译:持续学习旨在在非平稳数据流中学习,而不会严重遗忘以前的知识。目前的基于重播的方法通过使用一个保存已看到的数据的小缓冲区进行排练来解决这个挑战,这需要一个精细的样本选择策略。然而,现有的选择方案通常仅寻求最大化进行性选择的效用,忽略了连续选择之间的干扰。出于这个动机,我们在一个建立在影响函数上的框架中分析了顺序选择步骤的相互作用。我们设法确定了一类新的二阶影响,它将逐渐放大重播缓冲中的偶然偏差并破坏选择过程。为了正则化二阶影响,提出了一种新的选择目标,它还与两个广泛采用的标准有明显的联系。此外,我们提出了一种有效的实现方法来优化所提出的标准。在多个连续学习基准测试中的实验表明,我们的方法比现有方法具有优势。代码可在https://github.com/feifeiobama/InfluenceCL上获得。