A growing body of research in continual learning focuses on the catastrophic forgetting problem. While many attempts have been made to alleviate this problem, the majority of the methods assume a single model in the continual learning setup. In this work, we question this assumption and show that employing ensemble models can be a simple yet effective method to improve continual performance. However, the training and inference cost of ensembles can increase linearly with the number of models. Motivated by this limitation, we leverage the recent advances in the deep learning optimization literature, such as mode connectivity and neural network subspaces, to derive a new method that is both computationally advantageous and can outperform the state-of-the-art continual learning algorithms.
翻译:不断学习的越来越多的研究侧重于灾难性的遗忘问题。虽然为缓解这一问题做出了许多尝试,但大多数方法都假设了持续学习设置中的单一模式。在这项工作中,我们质疑这一假设,并表明采用共同模式可以是一个简单而有效的方法来提高持续业绩。然而,组合的培训和推论成本可以随着模型数量的增加而线性地增加。受这一限制的驱动,我们利用了在深度学习优化文献(如模式连接和神经网络子空间)方面的最新进展,以获得一种既在计算上有利又能超过最新持续学习算法的新方法。