In most machine learning algorithms, training data is assumed to be independent and identically distributed (iid). When it is not the case, the algorithm's performances are challenged, leading to the famous phenomenon of catastrophic forgetting. Algorithms dealing with it are gathered in the Continual Learning research field. In this paper, we study the regularization based approaches to continual learning and show that those approaches can not learn to discriminate classes from different tasks in an elemental continual benchmark: the class-incremental scenario. We make theoretical reasoning to prove this shortcoming and illustrate it with examples and experiments. Moreover, we show that it can have some important consequences on continual multi-tasks reinforcement learning or in pre-trained models used for continual learning. We believe that highlighting and understanding the shortcomings of regularization strategies will help us to use them more efficiently.
翻译:在大多数机学算法中,培训数据假定是独立的,而且分布相同(二)。如果不是这种情况,算法的性能就会受到挑战,导致著名的灾难性遗忘现象。它所涉及的等级分解方法收集在不断学习的研究领域。在本文中,我们研究基于正规化的不断学习方法,并表明这些方法不能在元素持续的基准中,即阶级增长情景中,区分不同任务类别。我们用理论推理来证明这一缺陷,并以实例和实验来说明这一点。此外,我们还表明,它可能对持续开展多任务强化学习或用于持续学习的预培训模式产生一些重要后果。我们认为,强调和理解正规化战略的缺陷将有助于我们更有效地利用这些缺陷。