Language models have become increasingly popular in recent years for tasks like information retrieval. As use-cases become oriented toward specific domains, fine-tuning becomes default for standard performance. To fine-tune these models for specific tasks and datasets, it is necessary to carefully tune the model's hyperparameters and training techniques. In this paper, we present an in-depth analysis of the performance of four transformer-based language models on the task of biomedical information retrieval. The models we consider are DeepMind's RETRO (7B parameters), GPT-J (6B parameters), GPT-3 (175B parameters), and BLOOM (176B parameters). We compare their performance on the basis of relevance, accuracy, and interpretability, using a large corpus of 480000 research papers on protein structure/function prediction as our dataset. Our findings suggest that smaller models, with <10B parameters and fine-tuned on domain-specific datasets, tend to outperform larger language models on highly specific questions in terms of accuracy, relevancy, and interpretability by a significant margin (+50% on average). However, larger models do provide generally better results on broader prompts.
翻译:近年来,对于信息检索等任务,语言模型越来越受欢迎。随着使用情况面向特定领域,微调就成为标准性能的默认。为微调这些模型对具体任务和数据集进行微调,有必要仔细调整模型的超参数和培训技术。在本文中,我们深入分析了四种基于变压器的语言模型在生物医学信息检索任务方面的性能。我们认为,这些模型是DeepMind的RETRO(7B参数)、GPT-J(6B参数)、GPT-3(175B参数)和BLOOM(176B参数),我们根据相关性、准确性和可解释性来比较这些模型的性能,我们用480 000份关于蛋白质结构/功能预测的大型研究论文作为数据集。我们的研究结果表明,规模较小的模型,加上<10B参数,并精细调了具体领域的数据集,往往在精确性、相关性和可解释性方面超越较大的语言模型(平均为+50%)。但是,较大的模型确实提供了更为广泛的及时的结果。