The aggregation of knowledge embedded in large language models (LLMs) holds the promise of new solutions to problems of observability and measurement in the social sciences. We examine this potential in a challenging setting: measuring latent ideology -- crucial for better understanding core political functions such as democratic representation. We scale pairwise liberal-conservative comparisons between members of the 116th U.S. Senate using prompts made to ChatGPT. Our measure strongly correlates with widely used liberal-conservative scales such as DW-NOMINATE. Our scale also has interpretative advantages, such as not placing senators who vote against their party for ideologically extreme reasons towards the middle. Our measure is more strongly associated with political activists' perceptions of senators than other measures, consistent with LLMs synthesizing vast amounts of politically relevant data from internet/book corpora rather than memorizing existing measures. LLMs will likely open new avenues for measuring latent constructs utilizing modeled information from massive text corpora.
翻译:聚合大型语言模型中嵌入的知识可以为社会科学中的观察和测量问题提供新的解决方案。我们研究了一个具有挑战性的设置:使用ChatGPT的提示来衡量潜在的意识形态-这对于更好地理解基本政治功能(如民主代表)至关重要。我们使用ChatGPT的提示对美国参议院116届成员之间的自由派-保守派两两比较进行了扩展。我们的测量与广泛使用的自由派-保守派刻度(如DW-NOMINATE)强烈相关。我们的比例也具有解释优势,例如不会将出于意识形态极端原因而反对所属党派的参议员置于中间。我们的测量与政治活动家对参议员的看法更密切相关,符合LLM从互联网/书籍语料库中综合大量与政治有关的数据而不是记忆现有测量的趋势。LLMs很可能为测量利用大文本语料库中建模信息的潜在构造打开新的途径。