Large pretrained language models have been performing increasingly well in a variety of downstream tasks via prompting. However, it remains unclear from where the model learns the task-specific knowledge, especially in a zero-shot setup. In this work, we want to find evidence of the model's task-specific competence from pretraining and are specifically interested in locating a very small subset of pretraining data that directly supports the model in the task. We call such a subset supporting data evidence and propose a novel method ORCA to effectively identify it, by iteratively using gradient information related to the downstream task. This supporting data evidence offers interesting insights about the prompted language models: in the tasks of sentiment analysis and textual entailment, BERT shows a substantial reliance on BookCorpus, the smaller corpus of BERT's two pretraining corpora, as well as on pretraining examples that mask out synonyms to the task verbalizers.
翻译:在一系列下游任务中,通过激励,大型的预先培训语言模型表现得越来越好;然而,从模型从何处学习具体任务的知识,特别是在零点设置中,情况仍然不清楚;在这项工作中,我们希望从预培训中找到模型具体任务能力的证据,并特别有兴趣找到能直接支持该模型的非常小的一组培训前数据。我们称之为这样一个子类数据支持证据,并提议一种新型的ORCA方法,通过迭接使用与下游任务有关的梯度信息来有效识别它。这种辅助数据证据为激发的语言模型提供了有趣的洞察力:在情绪分析和文字要求的任务中,BERT展示了对BookCorpus的极大依赖,这是BERT的两个小型的预培训公司,以及将同义词隐藏到任务语言的预培训范例。