The recent advent of large language models - large neural networks trained on a simple predictive objective over a massive corpus of natural language - has reinvigorated debate over whether human cognitive capacities might emerge in such generic models given sufficient training data. Of particular interest is the ability of these models to reason about novel problems zero-shot, without any direct training on those problems. In human cognition, this capacity is closely tied to an ability to reason by analogy. Here, we performed a direct comparison between human reasoners and a large language model (GPT-3) on a range of analogical tasks, including a novel text-based matrix reasoning task closely modeled on Raven's Progressive Matrices. We found that GPT-3 displayed a surprisingly strong capacity for abstract pattern induction, matching or even surpassing human capabilities in most settings. Our results indicate that large language models such as GPT-3 have acquired an emergent ability to find zero-shot solutions to a broad range of analogy problems.
翻译:最近大型语言模型的出现----大型神经网络在对大量自然语言进行简单预测目标方面受过培训的大型神经网络----使关于人类认知能力是否会在这类通用模型中出现的辩论重新活跃起来,因为有了足够的培训数据。特别令人感兴趣的是这些模型是否有能力在没有直接培训的情况下解释新的零弹问题。在人类认知中,这种能力与理性比喻的能力密切相关。在这里,我们直接比较了人类理性者和一个大型语言模型(GPT-3)关于一系列类比任务,包括一个基于文本的新颖的矩阵推理任务,它以雷文的进步矩阵为模型。我们发现,GPT-3在多数情况下,在抽象模式上岗、配对甚至超过人类能力方面表现出出奇异的强大能力。我们的结果表明,像GPT-3这样的大型语言模型已经获得了为广泛的类比问题找到零弹道解决方案的突然能力。