We study the problem of differentially private (DP) fine-tuning of large pre-trained models -- a recent privacy-preserving approach suitable for solving downstream tasks with sensitive data. Existing work has demonstrated that high accuracy is possible under strong privacy constraint, yet requires significant computational overhead or modifications to the network architecture. We propose differentially private bias-term fine-tuning (DP-BiTFiT), which matches the state-of-the-art accuracy for DP algorithms and the efficiency of the standard BiTFiT. DP-BiTFiT is model agnostic (not modifying the network architecture), parameter efficient (only training about $0.1\%$ of the parameters), and computation efficient (almost removing the overhead caused by DP, in both the time and space complexity). On a wide range of tasks, DP-BiTFiT is $2\sim 30\times$ faster and uses $2\sim 8\times$ less memory than DP full fine-tuning, even faster than the standard full fine-tuning. This amazing efficiency enables us to conduct DP fine-tuning on language and vision tasks with long-sequence texts and high-resolution images, which were computationally difficult using existing methods.
翻译:我们研究了对大型预先培训的模型进行差别私人(DP)微调的问题 -- -- 这是一种适合用敏感数据解决下游任务的最新隐私保护方法。现有工作表明,在强大的隐私限制下,高度精度是可能的,但需要大量的计算间接费用或网络结构的修改。我们提议对私人偏差性微调(DP-BitfiT)进行差别私人微调(DP-BitfiT),这种微调符合DP算法的最新精确度和BithiT标准的效率。DP-BitfiT是模型学问知(不改变网络结构)、参数效率(仅培训0.1美元)和计算效率(在时间和空间复杂度方面,基本上消除DP造成的间接费用)。关于广泛的任务,DP-BitfiT为2\sim 30美元,比DP全微调少记忆2\sim 8-tims$,甚至比标准全面微调快。这种惊人的效率使我们能够对语言和视觉任务进行微调,以长期的文本和高分辨率图像进行微调,这些是难以计算的方法。