With the advent of strong pre-trained natural language processing models like BERT, DeBERTa, MiniLM, T5, the data requirement for industries to fine-tune these models to their niche use cases has drastically reduced (typically to a few hundred annotated samples for achieving a reasonable performance). However, the availability of even a few hundred annotated samples may not always be guaranteed in low resource domains like automotive, which often limits the usage of such deep learning models in an industrial setting. In this paper we aim to address the challenge of fine-tuning such pre-trained models with only a few annotated samples, also known as Few-shot learning. Our experiments focus on evaluating the performance of a diverse set of algorithms and methodologies to achieve the task of classifying BOSCH automotive domain textual software requirements into 3 categories, while utilizing only 15 annotated samples per category for fine-tuning. We find that while SciBERT and DeBERTa based models tend to be the most accurate at 15 training samples, their performance improvement scales minimally as the number of annotated samples is increased to 50 in comparison to Siamese and T5 based models.
翻译:随着诸如BERT、DeBERTA、MiniLM、T5等经过预先训练的强力自然语言处理模型的出现,各行业对这些模型进行微调以适应其特殊用途案例的数据要求已大大减少(通常为达到合理性能而向数百个附加说明的样本);然而,在诸如汽车等低资源领域,甚至甚至几百个附加说明的样本也不一定总能得到保证,这往往限制了这种深层次学习模型在工业环境中的使用;在本文件中,我们旨在应对微调这类经过预先训练的模型的挑战,只有几个附加说明的样本,也称为少见的学习;我们实验的重点是评估一套不同的算法和方法的性能,以便完成将BOSCH汽车域域名软件要求分为三类的任务,同时只使用每类15个附加说明的样本进行微调;我们发现,虽然SciBERT和DeBERTA模型往往在15个培训样本中最准确,但其性能改进尺度最小,因为附加说明的样本数量与Siamese和T5模型相比增加到50个。