Recent work (e.g. LAMA (Petroni et al., 2019)) has found that the quality of the factual information extracted from Large Language Models (LLMs) depends on the prompts used to query them. This inconsistency is problematic because different users will query LLMs for the same information using different wording, but should receive the same, accurate responses regardless. In this work we aim to address this shortcoming by introducing P-Adapters: lightweight models that sit between the embedding layer and first attention layer of LLMs. They take LLM embeddings as input and output continuous prompts that are used to query the LLM. Additionally, we investigate Mixture of Experts (MoE) models that learn a set of continuous prompts ("experts") and select one to query the LLM. They require a separate classifier trained on human-annotated data to map natural language prompts to the continuous ones. P-Adapters perform comparably to the more complex MoE models in extracting factual information from BERT and RoBERTa while eliminating the need for additional annotations. P-Adapters show between 12-26% absolute improvement in precision and 36-50% absolute improvement in consistency over a baseline of only using natural language queries. Finally, we investigate what makes a P-adapter successful and conclude that access to the LLM's embeddings of the original natural language prompt, particularly the subject of the entity pair being asked about, is a significant factor.
翻译:最近的工作(例如LAMA(Petroni等人,2019年))发现,从大语言模型(LLMM)中提取的事实信息的质量取决于用于查询这些模型的提示。这种不一致问题在于,不同的用户会用不同的措辞询问LLMS的相同信息,但应该收到同样准确的答复。在这项工作中,我们的目标是通过引入P-Adapters:位于LLMS嵌入层和第一关注层之间的轻量级模型来解决这一缺陷。它们将LLMM嵌入作为投入和产出连续提示,用于查询LLM。此外,我们调查专家混合模型,以学习一套连续提示(“专家”)的模式,并选择一种用于查询LMM的相同信息。他们需要用一个单独的人类附加说明的数据来绘制自然语言的分类师,以更复杂的模式从BERT和RoBERTA中提取事实信息,同时消除关于LMMLM额外说明的需要。此外,我们调查专家混合的混合混合混合混合语言的精确度和精确度,在我们最后的精确度中,P-ADMLM的绝对精确度的精确度中,只有一个精确度的精确度的精确度的精确度的精确度,我们对等的精确度的精确度的精确度的精确度,最后的精确度是我们的精确度。