Differentially private stochastic gradient descent (DP-SGD) is the workhorse algorithm for recent advances in private deep learning. It provides a single privacy guarantee to all datapoints in the dataset. We propose an efficient algorithm to compute privacy guarantees for individual examples when releasing models trained by DP-SGD. We use our algorithm to investigate individual privacy parameters across a number of datasets. We find that most examples enjoy stronger privacy guarantees than the worst-case bound. We further discover that the training loss and the privacy parameter of an example are well-correlated. This implies groups that are underserved in terms of model utility are simultaneously underserved in terms of privacy guarantee. For example, on CIFAR-10, the average $\epsilon$ of the class with the lowest test accuracy is 26.3% higher than that of the class with the highest accuracy. We also run membership inference attacks to show this reflects disparate empirical privacy risks.
翻译:不同的私人静态梯度下降(DP-SGD)是私人深层学习最新进展的工作马算法(DP-SGD),它为数据集中的所有数据点提供了单一的隐私保障。我们提出一个高效的算法,在释放DP-SGD所培训的模型时计算个人案例的隐私保障。我们用我们的算法调查若干数据集的个人隐私参数。我们发现,大多数例子的隐私保障比最坏的例子更强。我们进一步发现,一个例子的训练损失和隐私参数是密切相关的。这意味着在模型效用方面服务不足的群体在隐私保障方面同时得不到充分服务。例如,在CIFAR-10中,测试精度最低的班级平均$\ epslon$比班级高26.3%。我们还进行成员推论攻击,以显示这种差异性的经验隐私风险。