Large-scale pre-trained language models (PLMs) are well-known for being capable of solving a task simply by conditioning a few input-label pairs dubbed demonstrations on a prompt without being explicitly tuned for the desired downstream task. Such a process (i.e., in-context learning), however, naturally leads to high reliance on the demonstrations which are usually selected from external datasets. In this paper, we propose self-generated in-context learning (SG-ICL), which generates demonstrations for in-context learning from PLM itself to minimize the reliance on the external demonstration. We conduct experiments on four different text classification tasks and show SG-ICL significantly outperforms zero-shot learning and is generally worth approximately 0.6 gold training samples. Moreover, our generated demonstrations show more consistent performance with low variance compared to randomly selected demonstrations from the training dataset.
翻译:众所周知,大规模预先培训的语言模型(PLMs)能够解决一项任务,只需将几对称为输入标签的演示配对设定为即时即刻的,而无需明确调整为预期的下游任务即可。然而,这一过程(即同文学习)自然导致高度依赖通常从外部数据集中挑选出来的演示。在本文中,我们提议自发的文字学习(SG-ICL),这会产生从PLM本身进行内文学习的演示,以尽量减少对外部演示的依赖。我们在四种不同的文本分类任务上进行实验,并显示SG-ICL大大超过零光学习,一般价值约为0.6个黄金培训样本。此外,我们制作的演示显示,与从培训数据集随机选定的演示相比,其表现更加一致,差异较小。