Fine-tuning pretrained language models (LMs) without making any architectural changes has become a norm for learning various language downstream tasks. However, for non-language downstream tasks, a common practice is to employ task-specific designs for input, output layers, and loss functions. For instance, it is possible to fine-tune an LM into an MNIST classifier by replacing the word embedding layer with an image patch embedding layer, the word token output layer with a 10-way output layer, and the word prediction loss with a 10-way classification loss, respectively. A natural question arises: Can LM fine-tuning solve non-language downstream tasks without changing the model architecture or loss function? To answer this, we propose Language-Interfaced Fine-Tuning (LIFT) and study its efficacy and limitations by conducting an extensive empirical study on a suite of non-language classification and regression tasks. LIFT does not make any changes to the model architecture or loss function, and it solely relies on the natural language interface, enabling "no-code machine learning with LMs." We find that LIFT performs comparably well across a wide range of low-dimensional classification and regression tasks, matching the performances of the best baselines in many cases, especially for the classification tasks. We also report experimental results on the fundamental properties of LIFT, including inductive bias, robustness, and sample complexity. We also analyze the effect of pretraining on LIFT and a few properties/techniques specific to LIFT, e.g., context-aware learning via appropriate prompting, calibrated predictions, data generation, and two-stage fine-tuning. Our code is available at https://github.com/UW-Madison-Lee-Lab/LanguageInterfacedFineTuning.
翻译:在不做任何建筑变革的情况下,微调预先训练的语言模型(LMS)已成为学习各种语言下游任务的一个规范。然而,对于非语言下游任务,通常的做法是对输入、输出层和损失功能采用特定任务设计。例如,有可能将LMM微调成MIS分类器,将LMM微调成MNIST的分类器,将LMM微调成一个图像嵌入层,用10个方向输出层取代一字表示输出层,用10个方向分类损失来预测单词预测损失。自然产生的一个问题是:LMM微调能解决非语言下游任务,而不改变模型结构或损失函数功能。为了回答这个问题,我们建议对LIFT进行语言内部精细精细精细设计,通过对非语言分类和回归任务进行广泛的实验性研究,LFTAF对模型结构或损失功能进行任何改变,并且仅仅依靠自然语言界面界面界面的界面,使LMSMS进行“无码校正”的校正。我们发现,LFTFT在一系列的精细的精细的模型和基础分析中,也匹配的精细的精细的精细分析。