A growing body of research in continual learning focuses on the catastrophic forgetting problem. While many attempts have been made to alleviate this problem, the majority of the methods assume a single model in the continual learning setup. In this work, we question this assumption and show that employing ensemble models can be a simple yet effective method to improve continual performance. However, ensembles' training and inference costs can increase significantly as the number of models grows. Motivated by this limitation, we study different ensemble models to understand their benefits and drawbacks in continual learning scenarios. Finally, to overcome the high compute cost of ensembles, we leverage recent advances in neural network subspace to propose a computationally cheap algorithm with similar runtime to a single model yet enjoying the performance benefits of ensembles.
翻译:越来越多的持续学习研究侧重于灾难性的遗忘问题。虽然为缓解这一问题做出了许多尝试,但大多数方法都假设了持续学习设置中的单一模式。在这项工作中,我们质疑这一假设,并表明采用混合模型可以是一个简单而有效的方法来提高连续性能。然而,组合的培训和推论成本可以随着模型数量的增加而大幅增加。受这一限制的驱动,我们研究不同的组合模型,以了解这些模型在持续学习的情景中的好处和缺点。最后,为了克服昆虫组的高计算成本,我们利用神经网络子空间最近的进展来提出一种计算成本低的算法,其运行时间与单一模型相似,但享受到合成的性能效益。