Differentially private stochastic gradient descent (DPSGD) is a variation of stochastic gradient descent based on the Differential Privacy (DP) paradigm, which can mitigate privacy threats that arise from the presence of sensitive information in training data. However, one major drawback of training deep neural networks with DPSGD is a reduction in the models accuracy. In this paper, we study the effect of normalization layers on the performance of DPSGD. We demonstrate that normalization layers significantly impact the utility of deep neural networks with noisy parameters and should be considered essential ingredients of training with DPSGD. In particular, we propose a novel method for integrating batch normalization with DPSGD without incurring an additional privacy loss. With our approach, we are able to train deeper networks and achieve a better utility-privacy trade-off.
翻译:不同的私人悬浮梯度下降(DPSGD)是基于差异隐私模式的随机梯度下降(DPSGD)的一种变化,这可以减轻培训数据中敏感信息的存在所产生的隐私威胁,然而,与DPSGD一起培训深层神经网络的一个主要缺点是模型准确性下降。我们在本文件中研究了正常化层对DPSGD绩效的影响。我们证明,正常化层对具有噪音参数的深层神经网络的实用性产生了重大影响,应被视为与DPSGD培训的重要内容。特别是,我们提出了在不造成额外隐私损失的情况下将批量正常化与DPSGD相结合的新方法。我们的方法可以更深入地培训网络,实现更好的使用专利权交易。