Lexicon information and pre-trained models, such as BERT, have been combined to explore Chinese sequence labelling tasks due to their respective strengths. However, existing methods solely fuse lexicon features via a shallow and random initialized sequence layer and do not integrate them into the bottom layers of BERT. In this paper, we propose Lexicon Enhanced BERT (LEBERT) for Chinese sequence labelling, which integrates external lexicon knowledge into BERT layers directly by a Lexicon Adapter layer. Compared with the existing methods, our model facilitates deep lexicon knowledge fusion at the lower layers of BERT. Experiments on ten Chinese datasets of three tasks including Named Entity Recognition, Word Segmentation, and Part-of-Speech tagging, show that LEBERT achieves the state-of-the-art results.
翻译:由于各自的强项,已将词汇信息和事先培训的模型(如BERT)合并在一起,探索中国的顺序标签任务,但是,现有方法仅通过浅层和随机初始序列层将词汇特征引信化,而不将其纳入BERT的底层。在本文中,我们建议中国的顺序标签采用Lexicon Afried BERT(LEBERT)(LEBERT)(LEBERT)(LEBERT)(将外部词汇知识直接纳入BERT层)。与现有方法相比,我们的模式为BERT下层的深层词汇知识融合提供了便利。在十个中国数据集上进行了实验,其中包括命名实体识别、文字分割和部分语音标记等三项任务。实验显示,LEBERT(LEBERT)取得了最新结果。