The in-context learning capabilities of LLMs like GPT-3 allow annotators to customize an LLM to their specific tasks with a small number of examples. However, users tend to include only the most obvious patterns when crafting examples, resulting in underspecified in-context functions that fall short on unseen cases. Further, it is hard to know when "enough" examples have been included even for known patterns. In this work, we present ScatterShot, an interactive system for building high-quality demonstration sets for in-context learning. ScatterShot iteratively slices unlabeled data into task-specific patterns, samples informative inputs from underexplored or not-yet-saturated slices in an active learning manner, and helps users label more efficiently with the help of an LLM and the current example set. In simulation studies on two text perturbation scenarios, ScatterShot sampling improves the resulting few-shot functions by 4-5 percentage points over random sampling, with less variance as more examples are added. In a user study, ScatterShot greatly helps users in covering different patterns in the input space and labeling in-context examples more efficiently, resulting in better in-context learning and less user effort.
翻译:GPT-3 等LLMs的内文字学习能力使批注者能够将LLM定制成其具体任务,并举几个例子。然而,用户往往在编织实例时只包括最明显的模式,造成未详细指定的内文字功能,而这种功能在不可见的案例中却不尽人意。此外,即使根据已知的模式,也很难知道何时才列入“足够的”实例。在这项工作中,我们介绍了ScatterShot,一个用于建立高质量的内文字学习演示集的交互式系统ScatterShot。ScatterShot反复地将未贴标签的数据切成任务特定模式,以积极的学习方式将探索不足或不饱和的切片中的信息输入样本进行抽样,并帮助用户在LLM和当前范例集的帮助下更有效地标注。在对两种文本扰动情景的模拟研究中,ScatterShot抽样使由此产生的微小功能比随机抽样高出4至5个百分点,而增加的例子。在用户研究中,StatterShot 大大帮助用户在空间投入和标签中学习更低的文本方面进行更好的学习。