The vanilla Differentially-Private Stochastic Gradient Descent (DP-SGD), including DP-Adam and other variants, ensures the privacy of training data by uniformly distributing privacy costs across training steps. The equivalent privacy costs controlled by maintaining the same gradient clipping thresholds and noise powers in each step result in unstable updates and a lower model accuracy when compared to the non-DP counterpart. In this paper, we propose the dynamic DP-SGD (along with dynamic DP-Adam, and others) to reduce the performance loss gap while maintaining privacy by dynamically adjusting clipping thresholds and noise powers while adhering to a total privacy budget constraint. Extensive experiments on a variety of deep learning tasks, including image classification, natural language processing, and federated learning, demonstrate that the proposed dynamic DP-SGD algorithm stabilizes updates and, as a result, significantly improves model accuracy in the strong privacy protection region when compared to the vanilla DP-SGD. We also conduct theoretical analysis to better understand the privacy-utility trade-off with dynamic DP-SGD, as well as to learn why Dynamic DP-SGD can outperform vanilla DP-SGD.
翻译:包括DP-Adam和其他变体在内的香草、差异型私人蒸汽梯底部(DP-SGD)通过在培训的各个阶段统一分配隐私费用,确保培训数据的隐私,通过保持同样的梯度剪切阈值和每个步骤的噪音力量来控制相等的隐私费用,导致与非DP对应方相比,动态DP-SGD的更新不稳定和模型准确性较低。在本文件中,我们提议动态DP-SGD(连同动态DP-Adam等)减少性能损失差距,同时通过动态调整剪切阈值和噪声能力,同时坚持完全的隐私预算限制,以保持隐私数据隐私。关于各种深层学习任务的广泛实验,包括图像分类、自然语言处理和联结学习,表明拟议的动态DP-SGD算法稳定了更新,并因此与香草DP-SGD相比,显著提高了强的隐私保护区的模型准确性。我们还进行理论分析,以更好地了解与动态DP-SGD的隐私权交易,并了解动态DP-SGD能否超越DP-SGD的范式。