Existing approaches to lifelong language learning rely on plenty of labeled data for learning a new task, which is hard to obtain in most real scenarios. Considering that humans can continually learn new tasks from a handful of examples, we expect the models also to be able to generalize well on new few-shot tasks without forgetting the previous ones. In this work, we define this more challenging yet practical problem as Lifelong Few-shot Language Learning (LFLL) and propose a unified framework for it based on prompt tuning of T5. Our framework called LFPT5 takes full advantage of PT's strong few-shot learning ability, and simultaneously trains the model as a task solver and a data generator. Before learning a new domain of the same task type, LFPT5 generates pseudo (labeled) samples of previously learned domains, and later gets trained on those samples to alleviate forgetting of previous knowledge as it learns the new domain. In addition, a KL divergence loss is minimized to achieve label consistency between the previous and the current model. While adapting to a new task type, LFPT5 includes and tunes additional prompt embeddings for the new task. With extensive experiments, we demonstrate that LFPT5 can be applied to various different types of tasks and significantly outperform previous methods in different LFLL settings.
翻译:终身语言学习的现有方法依靠大量标签数据来学习新任务,在多数真实情况下很难获得。考虑到人类可以不断从几个例子中学习新任务,我们期望模型也能在不忘记前几个例子的情况下对新的少发任务进行全面推广。在这项工作中,我们将这一更具挑战性和实际性的问题定义为终身少发语言学习(LFLLL),并提议一个基于迅速调控T5的统一框架。我们称为LFPT5的框架充分利用了PT的强力少发学习能力,同时将模型培训成任务解答器和数据生成器。在学习同一任务类型的新领域之前,LFPT5生成了以前所学域的假(标签)样本,后来又接受了这些样本的培训,以缓解在学习新领域时对先前知识的忘却。此外,为了在适应新任务类型的同时,LFPT5还包含和调整了额外的快速嵌入新任务。我们通过广泛实验,可以将不同的LF5 展示不同的LF形式应用不同的方法。