Prompt-based methods have been successfully applied in sentence-level few-shot learning tasks, mostly owing to the sophisticated design of templates and label words. However, when applied to token-level labeling tasks such as NER, it would be time-consuming to enumerate the template queries over all potential entity spans. In this work, we propose a more elegant method to reformulate NER tasks as LM problems without any templates. Specifically, we discard the template construction process while maintaining the word prediction paradigm of pre-training models to predict a class-related pivot word (or label word) at the entity position. Meanwhile, we also explore principled ways to automatically search for appropriate label words that the pre-trained models can easily adapt to. While avoiding complicated template-based process, the proposed LM objective also reduces the gap between different objectives used in pre-training and fine-tuning, thus it can better benefit the few-shot performance. Experimental results demonstrate the effectiveness of the proposed method over bert-tagger and template-based method under few-shot setting. Moreover, the decoding speed of the proposed method is up to 1930.12 times faster than the template-based method.
翻译:快速方法已成功地应用于判决一级少见的学习任务,这主要是由于模板和标签词词的复杂设计。然而,当适用于象征性标签任务,例如净化网络时,在所有潜在实体范围内列出模板查询将耗费时间。在这项工作中,我们建议采用一种更优雅的方法,将净化任务重新表述为没有模板的LM问题。具体地说,我们放弃模板构建过程,同时保持培训前模式的词预测范式,以便在实体位置上预测一个与阶级相关的主词(或标签单词)。同时,我们还探索原则性方法,自动寻找经过培训的模型容易适应的适当标签词。拟议的LM目标在避免基于模板的复杂进程的同时,还缩小了培训前和微调中使用的不同目标之间的差距,从而能够更好地使微调的性能受益。实验结果显示,拟议方法在几发式设置下比Birt-tagger和基于模板的方法的有效性。此外,拟议方法的解码速度比基于模板的方法快到190.12倍。