Differentially private stochastic gradient descent (DP-SGD) is the workhorse algorithm for recent advances in private deep learning. It provides a single privacy guarantee to all datapoints in the dataset. We propose output-specific $(\varepsilon,\delta)$-DP to characterize privacy guarantees for individual examples when releasing models trained by DP-SGD. We also design an efficient algorithm to investigate individual privacy across a number of datasets. We find that most examples enjoy stronger privacy guarantees than the worst-case bound. We further discover that the training loss and the privacy parameter of an example are well-correlated. This implies groups that are underserved in terms of model utility simultaneously experience weaker privacy guarantees. For example, on CIFAR-10, the average $\varepsilon$ of the class with the lowest test accuracy is 44.2% higher than that of the class with the highest accuracy.
翻译:差异性私人悬浮梯度下降(DP-SGD)是私人深层学习最新进展的工作马算法(DP-SGD),它为数据集中的所有数据点提供了单一的隐私保障。我们提议在释放DP-SGD培训的模型时,为单个例子提供特定产出的隐私保障($(varepsilon,\delta)$-DP)。我们还设计了一种有效的算法,以调查多个数据集的个人隐私。我们发现,大多数例子比最坏的数据集享有更强的隐私保障。我们进一步发现,一个示例的训练损失和隐私参数是完全相关联的。这意味着在模型效用方面服务不足的群体同时经历较弱的隐私保障。例如,在CIFAR-10上,测试精度最低的班的平均$(varepslon)比最精确的班级高出44.2%。