When learning tasks over time, artificial neural networks suffer from a problem known as Catastrophic Forgetting (CF). This happens when the weights of a network are overwritten during the training of a new task causing forgetting of old information. To address this issue, we propose MetA Reusable Knowledge or MARK, a new method that fosters weight reusability instead of overwriting when learning a new task. Specifically, MARK keeps a set of shared weights among tasks. We envision these shared weights as a common Knowledge Base (KB) that is not only used to learn new tasks, but also enriched with new knowledge as the model learns new tasks. Key components behind MARK are two-fold. On the one hand, a metalearning approach provides the key mechanism to incrementally enrich the KB with new knowledge and to foster weight reusability among tasks. On the other hand, a set of trainable masks provides the key mechanism to selectively choose from the KB relevant weights to solve each task. By using MARK, we achieve state of the art results in several popular benchmarks, surpassing the best performing methods in terms of average accuracy by over 10% on the 20-Split-MiniImageNet dataset, while achieving almost zero forgetfulness using 55% of the number of parameters. Furthermore, an ablation study provides evidence that, indeed, MARK is learning reusable knowledge that is selectively used by each task.
翻译:长期学习任务时, 人工神经网络会遇到被称为“ 灾难性遗忘” (CF) 的问题, 人工神经网络会遇到一个被称为“ 灾难性忘记” (CF) 的问题。 当一个网络的重量在培训新任务的过程中被超过时, 就会发生这种情况。 为了解决这个问题, 我们提出MetA可再利用的知识或MARK, 这是一种在学习新任务时促进重力再利用而不是超写的新方法。 具体地说, MARK 保持一套任务之间共有的权重。 我们把这些共享的权重设想为一个共同的知识库( KB), 它不仅用来学习新任务, 而且随着模型学习新任务而以新的知识丰富起来。 MARK 背后的关键组成部分是两重。 一方面, 金属化方法提供了一种关键机制, 以新的知识逐步丰富KB, 并促进任务之间的权重。 另一方面, 一组可重复的遮罩提供了从 KB 相关权重的权重。 我们通过使用MARK, 在若干通用基准中实现艺术成果状况, 超越了以平均的精确性方法, 以平均的精确度标准计算方法, 20- slegrelevildelexmlevilate 。