Prompt learning is a new paradigm in the Natural Language Processing (NLP) field which has shown impressive performance on a number of natural language tasks with common benchmarking text datasets in full, few-shot, and zero-shot train-evaluation setups. Recently, it has even been observed that large but frozen pre-trained language models (PLMs) with prompt learning outperform smaller but fine-tuned models. However, as with many recent NLP trends, the performance of even the largest PLMs such as GPT-3 do not perform well on specialized domains (e.g. medical text), and the common practice to achieve State of the Art (SoTA) results still consists of pre-training and fine-tuning the PLMs on downstream tasks. The reliance on fine-tuning large PLMs is problematic in clinical settings where data is often held in non-GPU environments, and more resource efficient methods of training specialized domain models is crucial. We investigated the viability of prompt learning on clinically meaningful decision tasks and directly compared with more traditional fine-tuning methods. Results are partially in line with the prompt learning literature, with prompt learning able to match or improve on traditional fine-tuning with substantially fewer trainable parameters and requiring less training data. We argue that prompt learning therefore provides lower computational resource costs applicable to clinical settings, that can serve as an alternative to fine-tuning ever increasing in size PLMs. Complementary code to reproduce experiments presented in this work can be found at: https://github.com/NtaylorOX/Public_Clinical_Prompt.
翻译:快速学习是自然语言处理(NLP)领域的新范例,它展示了许多自然语言任务的业绩,其共同基准文本数据集的完整、少发和零发的火车评价设置,在一些自然语言任务方面表现令人印象深刻。最近甚至发现,大型但冻结的预先培训语言模式(PLM),其迅速学习的超小型但微调的模式,与许多最近的自然语言处理(NLPP)趋势一样,甚至最大的PLM(如GPT-3)在专门领域(如医学文本)的绩效也不尽如人意,以及实现艺术状态(SoTA)成果的共同做法仍然包括培训前和对下游任务的PLMS进行微调。在临床环境中,对大PLMS的微调很困难,因为数据通常在非GPUP环境中保存,而培训专门领域模型的资源效率更高的方法也至关重要。我们调查了迅速学习具有临床意义的决定任务的可行性,直接与更传统的微调方法(如医学文本)的绩效与迅速学习文献是部分一致的,在快速学习能够匹配或改进传统的C-CTMS的临床环境,因此,对传统的快速进行更低的精确的校准,因此,对传统的CLMS进行更低的计算可以提供更低的计算。我们所需要的培训,因此,对传统的CLUDMSurbal-toal-drodu能能提供更低的计算。