Incremental Task learning (ITL) is a category of continual learning that seeks to train a single network for multiple tasks (one after another), where training data for each task is only available during the training of that task. Neural networks tend to forget older tasks when they are trained for the newer tasks; this property is often known as catastrophic forgetting. To address this issue, ITL methods use episodic memory, parameter regularization, masking and pruning, or extensible network structures. In this paper, we propose a new incremental task learning framework based on low-rank factorization. In particular, we represent the network weights for each layer as a linear combination of several rank-1 matrices. To update the network for a new task, we learn a rank-1 (or low-rank) matrix and add that to the weights of every layer. We also introduce an additional selector vector that assigns different weights to the low-rank matrices learned for the previous tasks. We show that our approach performs better than the current state-of-the-art methods in terms of accuracy and forgetting. Our method also offers better memory efficiency compared to episodic memory- and mask-based approaches. Our code will be available at https://github.com/CSIPlab/task-increment-rank-update.git
翻译:递增任务学习(ITL)是一个持续学习的类别,旨在为多重任务培训一个单一网络(一个又一个),每个任务的培训数据只有在培训时才能提供。神经网络在接受新任务培训时往往忘记旧任务;这种属性通常被称为灾难性的遗忘。为解决这一问题,国际交易日志的方法使用分层记忆、参数规范、遮盖和剪裁或扩展网络结构。在本文件中,我们提议了一个基于低等级因素化的新的递增任务学习框架。特别是,我们把每个层的网络重量作为若干一级矩阵的线性组合。为了更新新任务的网络,我们学习了一级(或低级)矩阵,并将其添加到每一层的重量中。我们还引入了另外一种选择矢量,对前一项任务所学的低级矩阵赋予不同重量。我们展示的方法在准确性和忘忘忘率方面比目前最先进的方法要好。我们的方法还提供更好的记忆效率,比前一级存储存储存储-存储-记录/Masmask-mask-maskas-maskasla) 方法。