Named entity recognition (NER) is a widely studied task in natural language processing. Recently, a growing number of studies have focused on the nested NER. The span-based methods, considering the entity recognition as a span classification task, can deal with nested entities naturally. But they suffer from the huge search space and the lack of interactions between entities. To address these issues, we propose a novel sequence-to-set neural network for nested NER. Instead of specifying candidate spans in advance, we provide a fixed set of learnable vectors to learn the patterns of the valuable spans. We utilize a non-autoregressive decoder to predict the final set of entities in one pass, in which we are able to capture dependencies between entities. Compared with the sequence-to-sequence method, our model is more suitable for such unordered recognition task as it is insensitive to the label order. In addition, we utilize the loss function based on bipartite matching to compute the overall training loss. Experimental results show that our proposed model achieves state-of-the-art on three nested NER corpora: ACE 2004, ACE 2005 and KBP 2017. The code is available at https://github.com/zqtan1024/sequence-to-set.
翻译:命名实体识别(NER)是自然语言处理中一项广泛研究的任务。最近,越来越多的研究侧重于嵌入式 NER 。基于光谱的方法,将实体识别视为一个跨级的分类任务,可以自然地与嵌入实体打交道。但是,它们受到巨大的搜索空间和实体之间缺乏互动的影响。为了解决这些问题,我们建议为嵌入的NER建立一个新型的序列到设置神经网络。我们不事先指定候选人的间隔,而是提供一套固定的可学习矢量,以了解有价值的宽度的模式。我们使用非不显性脱色器来预测最后一组实体,在其中我们能够捕捉到实体之间的依赖。与顺序到顺序方法相比,我们的模型更适合未经排序的识别任务,因为它对标签顺序敏感。此外,我们利用基于双方匹配的损失功能来计算总体培训损失。实验结果显示,我们提议的模型在三个嵌入式NER 102/KCONA/CEA CONAS 2017 上实现了“2005/CE” 和“ANCEACE-CAS-CO” 。