Prompting language models (LMs) with training examples and task descriptions has been seen as critical to recent successes in few-shot learning. In this work, we show that finetuning LMs in the few-shot setting can considerably reduce the need for prompt engineering. In fact, one can use null prompts, prompts that contain neither task-specific templates nor training examples, and achieve competitive accuracy to manually-tuned prompts across a wide range of tasks. While finetuning LMs does introduce new parameters for each downstream task, we show that this memory overhead can be substantially reduced: finetuning only the bias terms can achieve comparable or better accuracy than standard finetuning while only updating 0.1% of the parameters. All in all, we recommend finetuning LMs for few-shot learning as it is more accurate, robust to different prompts, and can be made nearly as efficient as using frozen LMs.
翻译:以培训实例和任务描述推动语言模式(LMS)已被视作是最近几张短片学习成功的关键。 在这项工作中,我们表明微调短片环境中的LMS可以大大降低快速工程的需求。 事实上,我们可以使用无效提示,既不含特定任务模板或培训范例的提示,也能够实现在广泛任务中手动调时的竞争性准确性。 微调LMS确实为每个下游任务引入了新的参数,但我们表明,这一记忆管理费用可以大幅降低:只微调偏差术语可以达到比标准微调更准确或更准确的精确度,而只更新0.1%的参数。 总之,我们建议微调LMS,因为其更准确性、对不同的速度更强,而且可以像使用冻结的LMS一样有效。