Prompt tuning is a parameter-efficient method, which freezes all PLM parameters and only prepends some additional tunable tokens called soft prompts to the input text. However, soft prompts heavily rely on a better initialization and may easily result in overfitting under few-shot settings, which causes prompt-tuning performing much worse than fine-tuning. To address the above issues, this paper proposes a novel Self-sUpervised Meta-prompt learning framework with MEtagradient Regularization for few shot generalization (SUMMER). We leverage self-supervised meta-learning to better initialize soft prompts and curriculum-based task augmentation is further proposed to enrich the meta-task distribution. Besides, a novel meta-gradient regularization method is integrated into the meta-prompt learning framework, which meta-learns to transform the raw gradient during few-shot learning into a domain-generalizable direction, thus alleviating the problem of overfitting. Extensive experiments show that SUMMER achieves better performance for different few-shot downstream tasks, and also exhibits a stronger domain generalization ability.
翻译:提示调整是一种参数高效的方法,它将所有PLM参数固定,仅在输入文本前添加一些附加的可调节标记,称为软提示。然而,软提示在很大程度上依赖于更好的初始化,并且在少样本设置下很容易导致过拟合,这导致提示调整比微调表现得更差。为了解决上述问题,本文提出了一种新颖的自监督元提示学习框架,其中包含MEtagradient正则化,用于少样本泛化(SUMMER)。我们利用自监督元学习更好地初始化软提示,并进一步提出基于课程的任务增强来丰富元任务分布。此外,新颖的元梯度正则化方法集成到元提示学习框架中,元学习将原始梯度转化为通用域方向,从而缓解过拟合问题。广泛的实验表明,SUMMER在不同的少样本下游任务中实现了更好的性能,并且还展现出更强的域泛化能力。