Active learning, which effectively collects informative unlabeled data for annotation, reduces the demand for labeled data. In this work, we propose to retrieve unlabeled samples with a local sensitivity and hardness-aware acquisition function. The proposed method generates data copies through local perturbations and selects data points whose predictive likelihoods diverge the most from their copies. We further empower our acquisition function by injecting the select-worst case perturbation. Our method achieves consistent gains over the commonly used active learning strategies in various classification tasks. Furthermore, we observe consistent improvements over the baselines on the study of prompt selection in prompt-based few-shot learning. These experiments demonstrate that our acquisition guided by local sensitivity and hardness can be effective and beneficial for many NLP tasks.
翻译:积极学习有效收集信息性、无标签的数据以作注释,减少了对标签数据的需求。在这项工作中,我们提议以本地敏感度和硬度认知获取功能检索未贴标签的样本。拟议方法通过本地扰动生成数据副本,并选择其预测可能性与副本相差最大的数据点。我们通过注射选定微弱案例扰动,进一步增强获取功能。我们的方法在各种分类任务中常用的积极学习战略上取得了一致的收益。此外,我们观察到,在快速选择短片学习的基线研究方面不断改进。这些实验表明,我们以本地敏感度和难度指导的获取能够对许多NLP任务有效且有益。