Prompt Tuning (PT) has been largely successful as a parameter-efficient way of conditioning large-scale pre-trained language models towards a downstream task. More recently, soft prompt tuning has aimed to learn a fixed set of task-specific continuous vectors, i.e., soft tokens that remain static across the task samples. However, a fixed prompt may not generalize well to the diverse kinds of inputs the task comprises. With this motivation, we propose a novel way of prompting, Vector-quantized Input-contextualized Prompt Tuning or VIP. Essentially, VIP focuses on two aspects i) input-adaptation: input-specific contextualization of the soft tokens; and ii) vector quantization: we pass the tokens through a quantizer which effectively reduces representation variance by sampling prompts from a compact latent space. Over a wide range of natural language understanding tasks (SuperGLUE, QA, Relation Classification, NER, NLI), our proposed VIP framework beats the PT model by a margin of 1.19\%. Additionally, on Out-of-domain QA and Multi-Task setups over 4 different tasks spanning over 12 domains, we find that VIP outperforms PT by 0.75\%.
翻译:快速调试(PT)在很大程度上作为一种节能的参数高效方式,使大规模预先培训的语言模式适应下游任务。最近,软快速调试旨在学习一套固定的任务特定连续矢量,即任务样品中保持静态的软象征物。然而,固定的快速调试可能没有很好地推广到任务所包含的各种投入。有了这一动机,我们提出了一种新型的促动方法,即矢量量化的输入-文字化快速调试或贵宾。基本上,贵宾侧重于两个方面:投入适应:一)投入特定背景化软象征物;二)矢量定量化:我们通过一个四分法通过一个四分法通过从一个紧凑的潜在空间取样来有效减少代表差异。在广泛的自然语言理解任务(SuperGLUE、QA、Relational 分类、NER、NLI)中,我们提议的贵宾框架比PT模式高出1.19分之差点。此外,在外方位QA和多式TFS-TA中,我们发现超过12个风险域。