Neural network models of language have long been used as a tool for developing hypotheses about conceptual representation in the mind and brain. For many years, such use involved extracting vector-space representations of words and using distances among these to predict or understand human behavior in various semantic tasks. In contemporary language AIs, however, it is possible to interrogate the latent structure of conceptual representations using methods nearly identical to those commonly used with human participants. The current work uses two common techniques borrowed from cognitive psychology to estimate and compare lexical-semantic structure in both humans and a well-known AI, the DaVinci variant of GPT-3. In humans, we show that conceptual structure is robust to differences in culture, language, and method of estimation. Structures estimated from AI behavior, while individually fairly consistent with those estimated from human behavior, depend much more upon the particular task used to generate behavior responses--responses generated by the very same model in the two tasks yield estimates of conceptual structure that cohere less with one another than do human structure estimates. The results suggest one important way that knowledge inhering in contemporary AIs can differ from human cognition.
翻译:神经网络语言模型长期以来一直被用作发展关于思维和大脑概念表示的假设工具。多年来,这种使用涉及提取单词的向量空间表示,并使用这些之间的距离来预测或理解各种语义任务中的人类行为。然而,在当今的语言人工智能中,可以使用与人类参与者几乎相同的方法来询问潜在的概念表示结构。当前的工作使用借鉴自认知心理学的两种常用技术来估计并比较人类和著名AI,即"GPT-3"的DaVinci变体的词汇语义结构。在人类中,我们表明概念结构对文化、语言和估计方法的差异是稳健的。从AI行为中估计出来的结构,虽然与从人类行为中估计出来的结构单独相当一致,但更多地取决于用于生成行为响应的特定任务——在两个任务中产生的响应所产生的概念结构估计与人类结构估计相比,更少地与之相互一致。这些结果表明,当代AI中内在的知识与人类认知有着重要的差异。