Large Language Models (LLMs) have become increasingly incorporated into everyday life for many internet users, taking on significant roles as advice givers in the domains of medicine, personal relationships, and even legal matters. The importance of these roles raise questions about how and what responses LLMs make in difficult political and moral domains, especially questions about possible biases. To quantify the nature of potential biases in LLMs, various works have applied Moral Foundations Theory (MFT), a framework that categorizes human moral reasoning into five dimensions: Harm, Fairness, Ingroup Loyalty, Authority, and Purity. Previous research has used the MFT to measure differences in human participants along political, national, and cultural lines. While there has been some analysis of the responses of LLM with respect to political stance in role-playing scenarios, no work so far has directly assessed the moral leanings in the LLM responses, nor have they connected LLM outputs with robust human data. In this paper we analyze the distinctions between LLM MFT responses and existing human research directly, investigating whether commonly available LLM responses demonstrate ideological leanings: either through their inherent responses, straightforward representations of political ideologies, or when responding from the perspectives of constructed human personas. We assess whether LLMs inherently generate responses that align more closely with one political ideology over another, and additionally examine how accurately LLMs can represent ideological perspectives through both explicit prompting and demographic-based role-playing. By systematically analyzing LLM behavior across these conditions and experiments, our study provides insight into the extent of political and demographic dependency in AI-generated responses.
翻译:大语言模型(LLMs)已日益融入众多互联网用户的日常生活,在医疗、人际关系甚至法律事务等领域承担着重要建议者的角色。这些角色的重要性引发了关于LLMs在复杂的政治与道德领域如何回应及回应内容的疑问,特别是对其潜在偏见的质疑。为量化LLMs中可能存在的偏见特性,已有研究采用道德基础理论(MFT)——该框架将人类道德推理划分为五个维度:伤害、公平、群体忠诚、权威和纯洁性。先前研究利用MFT测量了人类参与者在政治立场、国籍和文化背景上的差异。尽管已有研究在角色扮演场景中对LLMs回应与政治立场的关系进行了初步分析,但迄今尚未有工作直接评估LLM回应中的道德倾向,也未将LLM输出与可靠的人类数据进行关联。本文直接对比分析LLM的MFT回应与现有人类研究数据,探究常用LLM的回应是否呈现意识形态倾向:无论是通过其固有回应、对政治意识形态的直接表征,还是通过构建人类角色视角进行回应时的表现。我们评估LLMs是否固有地生成更贴近某一政治意识形态的回应,并进一步检验LLMs通过显式提示和基于人口统计学的角色扮演来表征意识形态视角的准确度。通过系统分析LLMs在不同条件和实验中的行为表现,本研究为理解AI生成回应中政治与人口统计学依赖的程度提供了新的见解。