We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-tuning of large pre-trained language models (PLMs) for few-shot learning. LiST improves over recent methods that adopt prompt-based fine-tuning (FN) using two key techniques. The first is the use of self-training to leverage large amounts of unlabeled data for prompt-based FN in few-shot settings. We use self-training in conjunction with meta-learning for re-weighting noisy pseudo-prompt labels. Self-training is expensive as it requires updating all the model parameters repetitively. Therefore, we use a second technique for light-weight fine-tuning where we introduce a small number of task-specific parameters that are fine-tuned during self-training while keeping the PLM encoder frozen. Our experiments show that LiST can effectively leverage unlabeled data to improve the model performance for few-shot learning. Additionally, the fine-tuning is efficient as it only updates a small percentage of parameters and the overall model footprint is reduced since several tasks can share a common PLM encoder as backbone. A comprehensive study on six NLU tasks demonstrate LiST to improve by 35% over classic fine-tuning and 6% over prompt-based FN with 96% reduction in number of trainable parameters when fine-tuned with no more than 30 labeled examples from each task. With only 14M tunable parameters, LiST outperforms GPT-3 in-context learning by 33% on few-shot NLU tasks.
翻译:我们为Lite Speed Astitution 提供了一种新的方法LiST, 用于Lite Special Development, 用于对大型预先训练的语言模型(PLMs)进行参数高效微调, 以进行少许学习。 自我训练费用昂贵, 因为它需要重复更新所有模型参数。 因此, 我们使用第二套轻度微调技术, 通过两种关键技术改进最近采用的方法, 采用基于即时微调(FN) 的快速微调(FN) 。 第一个是使用自我训练来利用大量无标签数据, 用于在短片环境中快速基于FNN的FN。 我们的实验显示, 将大量无标签数据有效地用于改善模拟学习的模型性能。 此外, 微调是有效的, 因为它只更新少量的参数和总体模型足迹, 因为一些任务可以重复更新所有模型参数。 因此, 我们使用第二套轻量的轻度微调微调微调微调微调技术, 在自我训练期间引入少量的低调低调的LL% 。 全面研究, 在最短调的LIST 任务中, 6级的LILIL 上, 上, 将每个LI-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx