Due to the high costs associated with finetuning large language models, various recent works propose to adapt them to specific tasks without any parameter updates through in-context learning. Unfortunately, for in-context learning there is currently no way to leverage unlabeled data, which is often much easier to obtain in large quantities than labeled examples. In this work, we therefore investigate ways to make use of unlabeled examples to improve the zero-shot performance of pretrained language models without any finetuning: We introduce Semantic-Oriented Unlabeled Priming (SOUP), a method that classifies examples by retrieving semantically similar unlabeled examples, assigning labels to them in a zero-shot fashion, and then using them for in-context learning. We also propose bag-of-contexts priming, a new priming strategy that is more suitable for our setting and enables the usage of more examples than fit into the context window.
翻译:由于与微调大型语言模型有关的高额成本,最近的各种工作都提议在不通过文字内学习更新参数的情况下,根据具体任务调整它们。 不幸的是,由于在文字内学习,目前没有办法利用未贴标签的数据,这通常比贴标签的例子容易得多。因此,我们研究如何利用未贴标签的例子来改进预先训练的语言模型的零弹性能,而不作任何微调:我们引入了语义-面向方向的未贴标签的普里明(SOUP)(SOUP)(SOUP)(SUP)(SUP)(SUP)(SUP))(SUP)(SUP)(SUP)(SUP)(SUP)(SUP)(SUP) (SUP) (SUP) (S) (SIM) (SIM) (SUP) (SUP) (SUP) (SUP) (S) (SUP) (SUP) (SIM) (S) (SIM) (SIM) (SIM) (SIM) (SIM) (SIM) (SIM) (SIM) (SIM) (SIM) (SIM) (SIM) (SUP) (S) (S) (IP) (SUP) (S) (S) (S) (IP) (S) (IP) (S) (S) (S) (S) (S) (S) (IP) (IP) (S) (S) (IP) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (S) (I) (I) (SU) (I) (S) (S) (I) (S) (S) (I) (IP) (IP) (I) (S) (S) (I) (S) (S) (I) (I) (I) (I) (I) (I) (I) (I) (