This paper reports on the evaluation of Deep Learning (DL) transformer architecture models for Named-Entity Recognition (NER) on ten low-resourced South African (SA) languages. In addition, these DL transformer models were compared to other Neural Network and Machine Learning (ML) NER models. The findings show that transformer models substantially improve performance when applying discrete fine-tuning parameters per language. Furthermore, fine-tuned transformer models outperform other neural network and machine learning models on NER with the low-resourced SA languages. For example, the transformer models obtained the highest F-scores for six of the ten SA languages and the highest average F-score surpassing the Conditional Random Fields ML model. Practical implications include developing high-performance NER capability with less effort and resource costs, potentially improving downstream NLP tasks such as Machine Translation (MT). Therefore, the application of DL transformer architecture models for NLP NER sequence tagging tasks on low-resourced SA languages is viable. Additional research could evaluate the more recent transformer architecture models on other Natural Language Processing tasks and applications, such as Phrase chunking, MT, and Part-of-Speech tagging.
翻译:本文报告了对南非10种低资源语言的命名-实体识别(NER)深层学习(DL)变压器结构模型的评价;此外,这些DL变压器模型与其他神经网络和机器学习(ML)NER模型进行了比较;结果显示,在应用每个语言的离散微调参数时,变压器模型大大改善了性能;此外,微调变压器模型优于其他神经网络和NER的机器学习模型;例如,10种南南非语言中,有6种语言的变压器模型获得了最高的F-核心,最高的平均F-核心模型超过了条件随机场模型;实际影响包括开发高性能净化能力,但精力和资源成本较低,有可能改进下游NLP的任务,如机器翻译(MT)。因此,将DLL变压器结构模型应用于NLP NER序列对低资源SA语言的标记任务是可行的。其他10种南南非语言中的6种的变压器结构模型,以及最高平均F-核心模型超过了条件随机场模型。实际影响包括以较低努力和资源成本的Phrage-MT,例如Phrage-GMTMT。