Prompts for pre-trained language models (PLMs) have shown remarkable performance by bridging the gap between pre-training tasks and various downstream tasks. Among these methods, prompt tuning, which freezes PLMs and only tunes soft prompts, provides an efficient and effective solution for adapting large-scale PLMs to downstream tasks. However, prompt tuning is yet to be fully explored. In our pilot experiments, we find that prompt tuning performs comparably with conventional full-model fine-tuning when downstream data are sufficient, whereas it performs much worse under few-shot learning settings, which may hinder the application of prompt tuning in practice. We attribute this low performance to the manner of initializing soft prompts. Therefore, in this work, we propose to pre-train prompts by adding soft prompts into the pre-training stage to obtain a better initialization. We name this Pre-trained Prompt Tuning framework "PPT". To ensure the generalization of PPT, we formulate similar classification tasks into a unified task form and pre-train soft prompts for this unified task. Extensive experiments show that tuning pre-trained prompts for downstream tasks can reach or even outperform full-model fine-tuning under both full-data and few-shot settings. Our approach is effective and efficient for using large-scale PLMs in practice.
翻译:培训前语言模型(PLMs)的提示通过缩小培训前任务和各种下游任务之间的差距,表现出了显著的成绩。在这些方法中,迅速调整(冻结PLMs)和只调软提示,为大规模PLMs适应下游任务提供了高效和有效的解决办法;然而,迅速调整还有待充分探索。在试点实验中,我们发现,当下游数据充足时,即迅速调整与常规全模微调相匹配,而下游数据充足时,其表现则不如常规全模微调,在少见的学习环境下则差得多,这可能会妨碍对实践的迅速调整。我们把这种低效归功于启动软提示的方式。因此,在这项工作中,我们建议通过在培训前阶段增加软提示来改进下游任务。我们称之为“PPTT”。为确保PPPT的普及,我们制定了类似的分类任务格式,为这一统一任务设计前的软动作,这可能会妨碍在实践中应用快速调整前的动作。我们提出的培训前的快速动作,在大型任务中,在升级后,在升级后,在升级后,在升级后,在全面进行。