While transfer learning has become a ubiquitous technique used across Natural Language Processing (NLP) tasks, it is often unable to replicate the performance of pre-trained models on text of niche domains like Automotive. In this paper we aim to understand the main characteristics of the distribution shift with automotive domain text (describing technical functionalities such as Cruise Control) and attempt to explain the potential reasons for the gap in performance. We focus on performing the Named Entity Recognition (NER) task as it requires strong lexical, syntactic and semantic understanding by the model. Our experiments with 2 different encoders, namely BERT-Base-Uncased and SciBERT-Base-Scivocab-Uncased have lead to interesting findings that showed: 1) The performance of SciBERT is better than BERT when used for automotive domain, 2) Fine-tuning the language models with automotive domain text did not make significant improvements to the NER performance, 3) The distribution shift is challenging as it is characterized by lack of repeating contexts, sparseness of entities, large number of Out-Of-Vocabulary (OOV) words and class overlap due to domain specific nuances.
翻译:虽然转让学习已成为在自然语言处理(NLP)任务中使用的无处不在的技术,但往往无法复制关于汽车等特殊领域文本的预先培训模型。在本文件中,我们的目标是通过汽车域文本(描述巡航控制等技术功能)来理解分销转换的主要特点,并试图解释造成绩效差距的潜在原因。我们侧重于执行命名实体识别(NER)任务,因为它需要该模型强有力的词汇、综合和语义理解。我们与两个不同的编码器(即BERT-Base-uncased和SciBERT-Base-Scivicocab-Uncase-Ucase)的实验得出了有趣的结论:(1) 用于汽车域文本时,SciBERT的性能优于BERT;(2) 使用汽车域文本对语言模型的微调没有显著改进NER的性能。(3) 分配变化具有挑战性,因为其特点是缺乏重复的环境、实体的分散性、大量外空间(OOVV)的字数和与特定域的类别重叠。