The strong few-shot in-context learning capability of large pre-trained language models (PLMs) such as GPT-3 is highly appealing for application domains such as biomedicine, which feature high and diverse demands of language technologies but also high data annotation costs. In this paper, we present the first systematic and comprehensive study to compare the few-shot performance of GPT-3 in-context learning with fine-tuning smaller (i.e., BERT-sized) PLMs on two highly representative biomedical information extraction tasks, named entity recognition and relation extraction. We follow the true few-shot setting to avoid overestimating models' few-shot performance by model selection over a large validation set. We also optimize GPT-3's performance with known techniques such as contextual calibration and dynamic in-context example retrieval. However, our results show that GPT-3 still significantly underperforms compared to simply fine-tuning a smaller PLM. In addition, GPT-3 in-context learning also yields smaller gains in accuracy when more training data becomes available. Our in-depth analyses further reveal issues of the in-context learning setting that may be detrimental to information extraction tasks in general. Given the high cost of experimenting with GPT-3, we hope our study provides guidance for biomedical researchers and practitioners towards more promising directions such as fine-tuning small PLMs.
翻译:GPT-3等经过预先训练的大型语言模型(PLMs)具有很强的点数内学习能力,这些语言模型具有很强的点数内学习能力,非常吸引生物医学等应用领域,这些应用领域具有语言技术的高要求和不同要求,但也具有高的数据注释成本。在本文件中,我们提出了第一份系统和全面的研究,将GPT-3的点数内学习成绩与小(即BERT大小)微调小(PLM)相比进行比较。此外,GPT-3的点数内学习在两项高度具有代表性的生物医学信息提取任务上,即名称实体识别和关系提取。我们遵循真正的点数设置,以避免在大型验证集中通过模型选择过高地估计模型的几分数性能。我们还以背景校准和动态内文本实例检索等已知技术优化GPT-3的性能。然而,我们的结果表明,GPT-3的性能学习成绩仍然大大低于微调低,而在更多的培训数据出现时,也会提高准确性能。我们深入的分析进一步揭示了精准性学习环境的问题,以便将GPLPT公司进行高额的实验。