Pre-trained Large Language Models (LLMs) are an integral part of modern AI that have led to breakthrough performances in complex AI tasks. Major AI companies with expensive infrastructures are able to develop and train these large models with billions and millions of parameters from scratch. Third parties, researchers, and practitioners are increasingly adopting these pre-trained models and fine-tuning them on their private data to accomplish their downstream AI tasks. However, it has been shown that an adversary can extract/reconstruct the exact training samples from these LLMs, which can lead to revealing personally identifiable information. The issue has raised deep concerns about the privacy of LLMs. Differential privacy (DP) provides a rigorous framework that allows adding noise in the process of training or fine-tuning LLMs such that extracting the training data becomes infeasible (i.e., with a cryptographically small success probability). While the theoretical privacy guarantees offered in most extant studies assume learning models from scratch through many training iterations in an asymptotic setting, this assumption does not hold in fine-tuning scenarios in which the number of training iterations is significantly smaller. To address the gap, we present \ewtune, a DP framework for fine-tuning LLMs based on Edgeworth accountant with finite-sample privacy guarantees. Our results across four well-established natural language understanding (NLU) tasks show that while \ewtune~adds privacy guarantees to LLM fine-tuning process, it directly contributes to decreasing the induced noise to up to 5.6\% and improves the state-of-the-art LLMs performance by up to 1.1\% across all NLU tasks. We have open-sourced our implementations for wide adoption and public testing purposes.
翻译:受过事先训练的大语言模型(LLMS)是现代AI(LLMS)的一个组成部分,它导致在复杂的AI任务中取得突破性业绩。具有昂贵基础设施的主要AI公司能够从零开始开发并培训出这些大型模型,其参数有数十亿甚至数百万个。第三方、研究人员和从业人员正在越来越多地采用这些经过预先训练的模型,并微调其私人数据以完成其下游的AI任务。然而,已经表明,对手可以提取/重新构建这些LLMS的精确培训样本,这可能导致披露个人可识别的信息。这个问题引起了对LLMS隐私的深切关注。 不同隐私(DP)提供了一个严格的框架,允许在培训或微调LMS的过程中增加噪音,从而提取培训数据变得不易(例如,在加密方面的可能性很小 ) 。虽然多数研究提供的理论隐私权保障假设通过在无深度环境中的许多培训来从抓取模型,但这一假设并不能通过微调国家对培训的RLMs(LMs)进行微调,在不断改进的LMSLMS-S-LLLLS-rodustringalalalalalal rodustring rodustring rodustring rodustris) 工作上,我们在不断改进我们的进度,在不断改进我们进行测试,在不断缩小的进度,在不断缩小的进度上显示我们的工作。