A significant shortcoming of current state-of-the-art (SOTA) named-entity recognition (NER) systems is their lack of generalization to unseen domains, which poses a major problem since obtaining labeled data for NER in a new domain is expensive and time-consuming. We propose ZERO, a model that performs zero-shot and few-shot learning in NER to generalize to unseen domains by incorporating pre-existing knowledge in the form of semantic word embeddings. ZERO first obtains contextualized word representations of input sentences using the model LUKE, reduces their dimensionality, and compares them directly with the embeddings of the external knowledge, allowing ZERO to be trained to recognize unseen output entities. We find that ZERO performs well on unseen NER domains with an average macro F1 score of 0.23, outperforms LUKE in few-shot learning, and even achieves competitive scores on an in-domain comparison. The performance across source-target domain pairs is shown to be inversely correlated with the pairs' KL divergence.
翻译:目前,最新技术(SOTA)命名实体识别(NER)系统的一个重大缺陷是,它们缺乏对隐蔽域的概括化,这是一个重大问题,因为在新领域获得标记的NER数据费用昂贵且耗时。我们提议ZERO模式,通过将原有知识以语义嵌入的形式纳入语义嵌入,将零点和几分学习结果推广到隐蔽域。ZERO首先利用LUKE模型获得输入句的背景化文字表达,降低其维度,并将之与外部知识的嵌入直接比较,使ZERO能够接受培训,以识别隐蔽输出实体。我们发现ZERO在隐蔽净化域上表现良好,平均宏观F1分为0.23分,在微小的学习中优于LUKE,在内部比较中甚至取得竞争性的评分。跨源目标域配对的性表现与两对KL差异呈反差。