Prompt-based learning, with its capability to tackle zero-shot and few-shot NLP tasks, has gained much attention in community. The main idea is to bridge the gap between NLP downstream tasks and language modeling (LM), by mapping these tasks into natural language prompts, which are then filled by pre-trained language models (PLMs). However, for prompt learning, there are still two salient gaps between NLP tasks and pretraining. First, prompt information is not necessarily sufficiently present during LM pretraining. Second, task-specific data are not necessarily well represented during pretraining. We address these two issues by proposing AdaPrompt, adaptively retrieving external data for continual pretraining of PLMs by making use of both task and prompt characteristics. In addition, we make use of knowledge in Natural Language Inference models for deriving adaptive verbalizers. Experimental results on five NLP benchmarks show that AdaPrompt can improve over standard PLMs in few-shot settings. In addition, in zero-shot settings, our method outperforms standard prompt-based methods by up to 26.35\% relative error reduction.
翻译:快速学习,由于有能力处理零射和几发NLP任务,在社区中引起了很大的注意。主要的想法是,通过将这些任务绘制成自然语言提示,然后由经过预先培训的语言模型(PLM)填补这些提示,从而缩小国家LP下游任务和语言模型(LM)之间的差距。然而,为了迅速学习,国家LP任务和预培训之间仍然有两个显著的差距。第一,在LM培训前阶段,迅速的信息不一定充分存在。第二,具体任务的数据不一定在培训前得到很好的反映。我们解决这两个问题的方法是:AdaPrompt, 适应性地检索外部数据,以便利用任务和迅速的特点对PLMS进行持续预先培训。此外,我们利用自然语言推断模型中的知识来产生适应性言语。五个国家LP基准的实验结果显示,AdaPrompt可以在少发环境中改进标准PLMs。此外,在零发环境中,我们的方法比标准快速方法高出26.35-相对错误减少。