Lifelong Language Learning (LLL) aims to train a neural network to learn a stream of NLP tasks while retaining knowledge from previous tasks. However, previous works which followed data-free constraint still suffer from catastrophic forgetting issue, where the model forgets what it just learned from previous tasks. In order to alleviate catastrophic forgetting, we propose the residual variational autoencoder (RVAE) to enhance LAMOL, a recent LLL model, by mapping different tasks into a limited unified semantic space. In this space, previous tasks are easy to be correct to their own distribution by pseudo samples. Furthermore, we propose an identity task to make the model is discriminative to recognize the sample belonging to which task. For training RVAE-LAMOL better, we propose a novel training scheme Alternate Lag Training. In the experiments, we test RVAE-LAMOL on permutations of three datasets from DecaNLP. The experimental results demonstrate that RVAE-LAMOL outperforms na\"ive LAMOL on all permutations and generates more meaningful pseudo-samples.
翻译:终身语言学习(LLLL)旨在培训神经网络,学习一串NLP任务,同时保留先前任务的知识。然而,以往在无数据限制之后开展的工作仍然受到灾难性的遗忘问题的影响,模型忘记了刚刚从以往任务中学到的东西。为了减轻灾难性的遗忘,我们建议残余变异自动编码器(RVAE)通过将不同任务映射到一个有限的统一语义空间,加强最近的LLLLL模型LOML。在这个空间,先前的任务很容易由伪样本来校正它们自己的分布。此外,我们提议让该模型具有识别属于哪个任务的样本的识别任务是歧视性的。为了更好地培训RVAE-LAMOL,我们提出了一个新的替代LAG培训计划。在实验中,我们测试RVAE-LAMOL对DecaNLP的三个数据集的配置。实验结果显示,RVAE-LAMOL在所有变形图中都存在“LAMOL”VAMOL 并产生更有意义的假样。