Fine-tuning continuous prompts for target tasks has recently emerged as a compact alternative to full model fine-tuning. Motivated by these promising results, we investigate the feasibility of extracting a discrete (textual) interpretation of continuous prompts that is faithful to the problem they solve. In practice, we observe a "wayward" behavior between the task solved by continuous prompts and their nearest neighbor discrete projections: We can find continuous prompts that solve a task while being projected to an arbitrary text (e.g., definition of a different or even a contradictory task), while being within a very small (2%) margin of the best continuous prompt of the same size for the task. We provide intuitions behind this odd and surprising behavior, as well as extensive empirical analyses quantifying the effect of various parameters. For instance, for larger model sizes we observe higher waywardness, i.e, we can find prompts that more closely map to any arbitrary text with a smaller drop in accuracy. These findings have important implications relating to the difficulty of faithfully interpreting continuous prompts and their generalization across models and tasks, providing guidance for future progress in prompting language models.
翻译:最近,对目标任务连续不断的微调作为全模型微调的缩略办法出现,作为替代全模型微调的缩略语。根据这些有希望的结果,我们调查了对忠实于所解决问题的连续快速进行独立(文字)解释的可行性。在实践中,我们观察到了连续快速解决的任务与近邻离散预测之间的“向上”行为:我们可以发现连续快速解决任务,同时被投向任意文本(例如,不同或甚至相互矛盾的任务的定义),同时处于同一任务最连续最迅速的幅度(2%)以内。我们提供了这种奇异和令人惊讶的行为背后的直觉,以及广泛的经验分析,对各种参数的效果进行了量化。例如,对于更大的模型规模,我们可以看到更接近任意文本的提示,但准确性较低。这些发现具有重要影响,即难以忠实地解释连续的迅速性以及这些模型和任务之间的概括性,为加速语言模型的未来进展提供指导。