Prompt tuning has been an extremely effective tool to adapt a pre-trained model to downstream tasks. However, standard prompt-based methods mainly consider the case of sufficient data of downstream tasks. It is still unclear whether the advantage can be transferred to the few-shot regime, where only limited data are available for each downstream task. Although some works have demonstrated the potential of prompt-tuning under the few-shot setting, the main stream methods via searching discrete prompts or tuning soft prompts with limited data are still very challenging. Through extensive empirical studies, we find that there is still a gap between prompt tuning and fully fine-tuning for few-shot learning. To bridge the gap, we propose a new prompt-tuning framework, called Soft Template Tuning (STT). STT combines manual and auto prompts, and treats downstream classification tasks as a masked language modeling task. Comprehensive evaluation on different settings suggests STT can close the gap between fine-tuning and prompt-based methods without introducing additional parameters. Significantly, it can even outperform the time- and resource-consuming fine-tuning method on sentiment classification tasks.
翻译:快速调整是使经过预先培训的模型适应下游任务的一个极为有效的工具,然而,标准的快速方法主要考虑下游任务的充分数据情况,目前还不清楚能否将优势转移给 " 点点数 " 制度,即每个下游任务只有有限的数据。虽然有些工作表明在 " 点数 " 设置下进行快速调整的潜力,但通过搜索离散提示或以有限数据调试软提示的主要流程方法仍然非常困难。通过广泛的实证研究,我们发现在快速调试和微调之间仍然存在着差距,为少发的学习进行充分微调。为弥合差距,我们提出了新的快速调控框架,称为 " 软模版调图案 " 。STT将人工和自动提示合并,并将下游分类任务作为隐蔽语言模式化任务处理。对不同环境的全面评价表明, " 点数 " 点数 " 可以在不引入额外参数的情况下弥合微调和快速调方法之间的差距。重要的是,它甚至超越了对情绪分类任务耗费时间和资源的调整方法。