Continual learning has become increasingly important as it enables NLP models to constantly learn and gain knowledge over time. Previous continual learning methods are mainly designed to preserve knowledge from previous tasks, without much emphasis on how to well generalize models to new tasks. In this work, we propose an information disentanglement based regularization method for continual learning on text classification. Our proposed method first disentangles text hidden spaces into representations that are generic to all tasks and representations specific to each individual task, and further regularizes these representations differently to better constrain the knowledge required to generalize. We also introduce two simple auxiliary tasks: next sentence prediction and task-id prediction, for learning better generic and specific representation spaces. Experiments conducted on large-scale benchmarks demonstrate the effectiveness of our method in continual text classification tasks with various sequences and lengths over state-of-the-art baselines. We have publicly released our code at https://github.com/GT-SALT/IDBR.
翻译:持续学习已变得日益重要,因为它使国家学习计划模式能够长期不断学习和获得知识,以往的不断学习方法主要是为了保存以往任务的知识,而没有大力强调如何将模式推广到新的任务中。在这项工作中,我们提议了一种基于信息分离的规范化方法,用于在文本分类方面不断学习。我们提出的方法首先将文本隐藏的空间分解为通用的表述形式,这些表述方式与每项任务特有的所有任务和表述形式都具有通用性,并进一步对这些表述方式进行不同的规范,以更好地限制一般化所需的知识。我们还引入了两项简单的辅助任务:下一句预测和任务化预测,以学习更好的通用和特定代表空间。在大规模基准上进行的实验显示了我们持续文本分类任务方法的效力,其顺序和长度超越了最新基线。我们在https://github.com/GT-SALT/IDBR公开发布了我们的代码。