The information bottleneck (IB) principle has been proven effective in various NLP applications. The existing work, however, only used either generative or information compression models to improve the performance of the target task. In this paper, we propose to combine the two types of IB models into one system to enhance Named Entity Recognition (NER). For one type of IB model, we incorporate two unsupervised generative components, span reconstruction and synonym generation, into a span-based NER system. The span reconstruction ensures that the contextualised span representation keeps the span information, while the synonym generation makes synonyms have similar representations even in different contexts. For the other type of IB model, we add a supervised IB layer that performs information compression into the system to preserve useful features for NER in the resulting span representations. Experiments on five different corpora indicate that jointly training both generative and information compression models can enhance the performance of the baseline span-based NER system. Our source code is publicly available at https://github.com/nguyennth/joint-ib-models.
翻译:信息瓶颈(IB)原则在各种NLP应用中证明是有效的,但是,现有的工作只使用了基因化或信息压缩模型来改进目标任务的业绩。在本文件中,我们提议将两种类型的IB模型合并为一个系统,以加强命名实体识别(NER)。对于一种IB模型,我们将两个未经监督的基因化组成部分,跨度重建和同义制生成,纳入一个基于跨系统的NER系统。这一重建确保了背景化的跨区域代表结构保持了信息,而同义地名一代甚至在不同的情况下也有类似的同义词表示。对于另一种IB模型,我们提议在系统中增加一个监督的IB层,对信息进行压缩,以便在由此形成的光谱表中保留NER的有用特征。对五种不同的公司进行的实验表明,对基因化和信息压缩模型的联合培训可以提高基于基线的NER系统的性能。我们的源代码公布在https://github.com/nguyenn/ interwoint-ib-models。