Domain-specific fine-tuning strategies for large pre-trained models received vast attention in recent years. In previously studied settings, the model architectures and parameters are tunable or at least visible, which we refer to as white-box settings. This work considers a new scenario, where we do not have access to a pre-trained model, except for its outputs given inputs, and we call this problem black-box fine-tuning. To illustrate our approach, we first introduce the black-box setting formally on text classification, where the pre-trained model is not only frozen but also invisible. We then propose our solution black-box prompt, a new technique in the prompt-learning family, which can leverage the knowledge learned by pre-trained models from the pre-training corpus. Our experiments demonstrate that the proposed method achieved the state-of-the-art performance on eight datasets. Further analyses on different human-designed objectives, prompt lengths, and intuitive explanations demonstrate the robustness and flexibility of our method.
翻译:近年来,大型预培训模式的具体微调战略受到广泛关注。在以往研究的环境下,模型架构和参数是可捕捉的,至少是可见的,我们称之为“白箱”设置。这项工作考虑了一种新的情景,我们除了提供的产出外,无法获得预培训模式,我们称之为“问题黑箱微调”。为了说明我们的方法,我们首先正式引入文本分类的黑箱设置,即预培训模式不仅被冻结,而且不为人知。然后,我们提出我们的解决方案黑箱提示,即快速学习家庭的一种新技术,可以利用预培训模式从培训前材料中获得的知识。我们的实验表明,拟议的方法在八个数据集上达到了最先进的表现。对不同的人类设计目标、提示长度和直觉解释的进一步分析显示了我们方法的稳健性和灵活性。