Probabilistic models of language understanding are interpretable and structured, for instance models of metaphor understanding describe inference about latent topics and features. However, these models are manually designed for a specific task. Large language models (LLMs) can perform many tasks through in-context learning, but they lack the clear structure of probabilistic models. In this paper, we use chain-of-thought prompts to introduce structures from probabilistic models into LLMs. These prompts lead the model to infer latent variables and reason about their relationships to choose appropriate paraphrases for metaphors. The latent variables and relationships chosen are informed by theories of metaphor understanding from cognitive psychology. We apply these prompts to the two largest versions of GPT-3 and show that they can improve paraphrase selection.
翻译:语言理解的概率模型是可以解释和结构化的,例如,隐喻理解模型描述潜在主题和特征的推论;然而,这些模型是手工设计的,用于具体任务;大型语言模型(LLMs)可以通过内通学习执行许多任务,但缺乏概率模型的明确结构。在本文中,我们利用思维链的提示将概率模型的结构引入LMs。这些提示引导模型推断潜在变量及其关系的理由,以选择隐喻的适当引言。所选择的潜在变量和关系以认知心理学的隐喻理解理论为依据。我们将这些提示应用于两种最大的GPT-3版本,并表明它们能够改进参数选择。