We propose and explore the possibility that language models can be studied as effective proxies for specific human sub-populations in social science research. Practical and research applications of artificial intelligence tools have sometimes been limited by problematic biases (such as racism or sexism), which are often treated as uniform properties of the models. We show that the "algorithmic bias" within one such tool -- the GPT-3 language model -- is instead both fine-grained and demographically correlated, meaning that proper conditioning will cause it to accurately emulate response distributions from a wide variety of human subgroups. We term this property "algorithmic fidelity" and explore its extent in GPT-3. We create "silicon samples" by conditioning the model on thousands of socio-demographic backstories from real human participants in multiple large surveys conducted in the United States. We then compare the silicon and human samples to demonstrate that the information contained in GPT-3 goes far beyond surface similarity. It is nuanced, multifaceted, and reflects the complex interplay between ideas, attitudes, and socio-cultural context that characterize human attitudes. We suggest that language models with sufficient algorithmic fidelity thus constitute a novel and powerful tool to advance understanding of humans and society across a variety of disciplines.
翻译:我们提议并探讨是否可以将语言模型作为社会科学研究中特定人类亚群群的有效替代物加以研究。人工智能工具的实际应用和研究应用有时受到问题偏见(例如种族主义或性别主义)的限制,这些偏见往往被视为模型的统一特性。我们表明,此类工具(GPT-3语言模型)中的“语言偏差”其实是细微和人口统计上相互关联的,这意味着适当的调节将促使它准确地模仿来自各种人类分组的反应分布。我们用GPT-3来形容这种属性“对等性”,并探索其范围。我们通过在美国进行的多次大型调查中对实际人类参与者的数千种社会-人口背层模型进行调整,从而创建“硅样本”。然后,我们比较“语言偏差”和人类样本,以证明GPT-3所含信息远远超出表面相似性。它是微妙、多面的,反映了作为人类态度特征的各种思想、态度和社会文化背景之间的复杂相互作用。我们建议,语言模型具有超越人类新式和先进性的新式,因此构成一种充分的人际关系和先进性。