Prompt tuning, in which a base pretrained model is adapted to each task via conditioning on learned prompt vectors, has emerged as a promising approach for efficiently adapting large language models to multiple downstream tasks. However, existing methods typically learn soft prompt vectors from scratch, and it has not been clear how to exploit the rich cross-task knowledge with prompt vectors in a multitask learning setting. We propose multitask prompt tuning (MPT), which first learns a single transferable prompt by distilling knowledge from multiple task-specific source prompts. We then learn multiplicative low rank updates to this shared prompt to efficiently adapt it to each downstream target task. Extensive experiments on 23 NLP datasets demonstrate that our proposed approach outperforms the state-of-the-art methods, including the full finetuning baseline in some cases, despite only tuning 0.035% as many task-specific parameters.
翻译:快速调试,在这种调试中,一个基础预先培训的模型通过对学习到的快速矢量进行调节,适应每项任务,已成为一种大语言模型有效适应多个下游任务的有希望的方法。然而,现有方法通常从零开始学习软性快速矢量,而且尚不清楚如何在多任务学习环境中利用丰富的跨任务知识,在多任务学习环境中利用快速矢量。我们提议多任务快速调试(MPT),它首先通过从多个任务特定源的提示中提取知识来学习单一可转让的可转让性。然后,我们学习多复制的低级别更新,从而快速共享,从而有效地适应每个下游目标任务。关于23个新任务数据集的广泛实验表明,我们拟议的方法超越了最新方法,包括在某些情况下的全面微调基准,尽管许多任务特定参数仅调出0.035%。</s>