Autoregressive language models, which use deep learning to produce human-like texts, have become increasingly widespread. Such models are powering popular virtual assistants in areas like smart health, finance, and autonomous driving. While the parameters of these large language models are improving, concerns persist that these models might not work equally for all subgroups in society. Despite growing discussions of AI fairness across disciplines, there lacks systemic metrics to assess what equity means in dialogue systems and how to engage different populations in the assessment loop. Grounded in theories of deliberative democracy and science and technology studies, this paper proposes an analytical framework for unpacking the meaning of equity in human-AI dialogues. Using this framework, we conducted an auditing study to examine how GPT-3 responded to different sub-populations on crucial science and social topics: climate change and the Black Lives Matter (BLM) movement. Our corpus consists of over 20,000 rounds of dialogues between GPT-3 and 3290 individuals who vary in gender, race and ethnicity, education level, English as a first language, and opinions toward the issues. We found a substantively worse user experience with GPT-3 among the opinion and the education minority subpopulations; however, these two groups achieved the largest knowledge gain, changing attitudes toward supporting BLM and climate change efforts after the chat. We traced these user experience divides to conversational differences and found that GPT-3 used more negative expressions when it responded to the education and opinion minority groups, compared to its responses to the majority groups. We discuss the implications of our findings for a deliberative conversational AI system that centralizes diversity, equity, and inclusion.
翻译:虽然这些大型语言模型的参数正在改善,但人们仍然担心这些模型可能不能平等地适用于社会上的所有分组。尽管关于不同学科的AI公平的讨论越来越多,但缺乏系统化的衡量标准来评估对话体系中的公平意味着什么以及如何让不同的人口参与评估循环。本文以审议民主理论和科学技术研究为基础,提出了一个分析框架,以解开人类-AI对话中的公平含义。我们利用这个框架,进行了一项中央审计研究,以研究GPT-3如何应对关键科学和社会问题的不同亚群:气候变化和黑人生活物质运动。我们的材料包括GPT-3和3 3 290个在性别、种族和族裔、教育程度、英语作为首选语言和观点上的差异。我们发现,GPT-3在支持人类-AI对话中的公平含义方面,用户经验要差得多。我们发现,GPT-3在支持少数群体观点和教育分群中是如何应对的。 与GPT-3相比,这两个群体在GL对话中发现,在GPT-L对话中,在用户对话中找到了最大的知识差异。</s>