The first edition of the IberLEF 2021 shared task on automatic detection of borrowings (ADoBo) focused on detecting lexical borrowings that appeared in the Spanish press and that have recently been imported into the Spanish language. In this work, we tested supplementary training on intermediate labeled-data tasks (STILTs) from part of speech (POS), named entity recognition (NER), code-switching, and language identification approaches to the classification of borrowings at the token level using existing pre-trained transformer-based language models. Our extensive experimental results suggest that STILTs do not provide any improvement over direct fine-tuning of multilingual models. However, multilingual models trained on small subsets of languages perform reasonably better than multilingual BERT but not as good as multilingual RoBERTa for the given dataset.
翻译:IberLEF 2021年IberLEF 关于自动检测借款的共同任务(ADoBo)的第一版侧重于发现西班牙报刊上出现并于最近进口到西班牙文的词汇借款,在这项工作中,我们测试了部分演讲(POS)、名称实体识别(NER)、编码转换和语言识别方法等中间标记数据任务的补充培训,以便利用现有预先培训的变压器变压器语言模式,在象征性水平上对借款进行分类。我们广泛的实验结果表明,科技创新技术在直接微调多语模式方面没有任何改进,然而,在小类语言上培训的多语模式比多语种BERT要好得多,但与给定数据集的多语种 RoBERTA相比不那么好。