Fine-tuning has proven to be highly effective in adapting pre-trained models to perform better on new desired tasks with minimal data samples. Among the most widely used approaches are reparameterization methods, which update a target module by augmenting its frozen weight matrix with an additional trainable weight matrix. The most prominent example is Low Rank Adaption (LoRA), which gained significant attention in recent years. In this paper, we introduce a new class of reparameterization methods for transfer learning, designed to enhance the generalization ability of fine-tuned models. We establish the effectiveness of our approach in a high-dimensional binary classification setting using tools from Random Matrix Theory, and further validate our theoretical findings through more realistic experiments, such as fine-tuning LLMs.
翻译:微调已被证明在使预训练模型以最少的数据样本在新任务上表现更佳方面极为有效。最广泛使用的方法之一是重参数化方法,该方法通过向冻结权重矩阵添加额外的可训练权重矩阵来更新目标模块。最突出的例子是低秩自适应(LoRA),近年来获得了广泛关注。本文提出了一类新的用于迁移学习的重参数化方法,旨在提升微调模型的泛化能力。我们利用随机矩阵理论工具,在高维二分类场景中验证了该方法的有效性,并通过更实际的实验(如大语言模型微调)进一步证实了理论结果。