In this paper, we move towards combining large parametric models with non-parametric prototypical networks. We propose prototypical fine-tuning, a novel prototypical framework for fine-tuning pretrained language models (LM), which automatically learns a bias to improve predictive performance for varying data sizes, especially low-resource settings. Our prototypical fine-tuning approach can automatically adjust the model capacity according to the number of data points and the model's inherent attributes. Moreover, we propose four principles for effective prototype fine-tuning towards the optimal solution. Experimental results across various datasets show that our work achieves significant performance improvements under various low-resource settings, as well as comparable and usually better performances in high-resource scenarios.
翻译:在本文中,我们开始将大型参数模型与非参数准典型网络结合起来。我们提出了原型微调,这是对预先培训的语言模型进行微调的新颖的原型典型框架,它自动学会一种偏差,可以提高不同数据大小的预测性能,特别是低资源环境的预测性能。我们的原型微调方法可以根据数据点的数量和模型的固有属性自动调整模型能力。此外,我们提出了有效调整原型的四项原则,以便实现最佳解决方案。各种数据集的实验结果表明,在各种低资源环境中,我们的工作取得了显著的绩效改进,高资源情景下的业绩也具有可比性,通常也比较好。