The emergence of large language models (LLMs) represents a major advance in artificial intelligence (AI) research. However, the widespread use of LLMs is also coupled with significant ethical and social challenges. Previous research has pointed towards auditing as a promising governance mechanism to help ensure that AI systems are designed and deployed in ways that are ethical, legal, and technically robust. However, existing auditing procedures fail to address the governance challenges posed by LLMs, which are adaptable to a wide range of downstream tasks. To help bridge that gap, we offer three contributions in this article. First, we establish the need to develop new auditing procedures that capture the risks posed by LLMs by analysing the affordances and constraints of existing auditing procedures. Second, we outline a blueprint to audit LLMs in feasible and effective ways by drawing on best practices from IT governance and system engineering. Specifically, we propose a three-layered approach, whereby governance audits, model audits, and application audits complement and inform each other. Finally, we discuss the limitations not only of our three-layered approach but also of the prospect of auditing LLMs at all. Ultimately, this article seeks to expand the methodological toolkit available to technology providers and policymakers who wish to analyse and evaluate LLMs from technical, ethical, and legal perspectives.
翻译:大型语言模型(LLMS)的出现是人工智能(AI)研究的一大进步,然而,广泛使用LLMS也伴随着重大的道德和社会挑战。以前的研究指出,审计是一个很有希望的治理机制,有助于确保AI系统的设计和应用方式符合道德、法律和技术上稳健。但是,现有的审计程序未能解决LLMS提出的治理挑战,因为LLMS适应广泛的下游任务。为了帮助弥合这一差距,我们在本条中提供了三项贡献。首先,我们确定需要制定新的审计程序,通过分析现有审计程序的费用负担和制约因素,抓住LLMS带来的风险。第二,我们通过借鉴信息技术治理和系统工程的最佳做法,勾勒订出可行和有效审计LMS的蓝图。具体地说,我们提出了三层方法,即治理审计、示范审计和应用审计相互补充和相互通报。最后,我们不仅讨论了我们三层方法的局限性,而且还讨论了审计LMS的前景。最后,我们试图从技术提供者、道德观点和决策者的角度,从技术角度,扩大现有方法工具包。