Nowadays, owners and developers of deep learning models must consider stringent privacy-preservation rules of their training data, usually crowd-sourced and retaining sensitive information. The most widely adopted method to enforce privacy guarantees of a deep learning model nowadays relies on optimization techniques enforcing differential privacy. According to the literature, this approach has proven to be a successful defence against several models' privacy attacks, but its downside is a substantial degradation of the models' performance. In this work, we compare the effectiveness of the differentially-private stochastic gradient descent (DP-SGD) algorithm against standard optimization practices with regularization techniques. We analyze the resulting models' utility, training performance, and the effectiveness of membership inference and model inversion attacks against the learned models. Finally, we discuss differential privacy's flaws and limits and empirically demonstrate the often superior privacy-preserving properties of dropout and l2-regularization.
翻译:目前,深层学习模式的拥有者和开发者必须考虑其培训数据的严格的隐私保护规则,通常是由众人提供和保留敏感信息。目前,最广泛采用的对深层学习模式实施隐私保障的方法依靠的是实施差异隐私的优化技术。根据文献,这种方法已证明是针对若干模式隐私攻击的成功防御,但其劣势是模型性能的大幅下降。在这项工作中,我们比较了差异私人随机梯度梯度(DP-SGD)算法与标准优化做法(DP-SGD)的实效和规范化技术。我们分析了由此产生的模型的实用性、培训绩效、成员推论的有效性以及针对所学模型的反向攻击。最后,我们讨论了差异隐私的缺陷和限制,从经验上证明了辍学和12个常规化的隐私保护特性往往更高。