The data used to train deep neural network (DNN) models in applications such as healthcare and finance typically contain sensitive information. A DNN model may suffer from overfitting. Overfitted models have been shown to be susceptible to query-based attacks such as membership inference attacks (MIAs). MIAs aim to determine whether a sample belongs to the dataset used to train a classifier (members) or not (nonmembers). Recently, a new class of label based MIAs (LAB MIAs) was proposed, where an adversary was only required to have knowledge of predicted labels of samples. Developing a defense against an adversary carrying out a LAB MIA on DNN models that cannot be retrained remains an open problem. We present LDL, a light weight defense against LAB MIAs. LDL works by constructing a high-dimensional sphere around queried samples such that the model decision is unchanged for (noisy) variants of the sample within the sphere. This sphere of label-invariance creates ambiguity and prevents a querying adversary from correctly determining whether a sample is a member or a nonmember. We analytically characterize the success rate of an adversary carrying out a LAB MIA when LDL is deployed, and show that the formulation is consistent with experimental observations. We evaluate LDL on seven datasets -- CIFAR-10, CIFAR-100, GTSRB, Face, Purchase, Location, and Texas -- with varying sizes of training data. All of these datasets have been used by SOTA LAB MIAs. Our experiments demonstrate that LDL reduces the success rate of an adversary carrying out a LAB MIA in each case. We empirically compare LDL with defenses against LAB MIAs that require retraining of DNN models, and show that LDL performs favorably despite not needing to retrain the DNNs.
翻译:用于在医疗保健和融资等应用中培养深神经网络(DNN)模型的数据通常包含敏感信息。 DNN模式可能受到过度改造。 超称模型被证明容易受到基于质询的攻击, 如会籍推断攻击(MIAs)。 MIA旨在确定样本是否属于用于培训分类员(成员)或非成员(非成员)的数据集。 最近, 提议了一个新的标签类别, 以MIA( LAB MIAs) 为基础的模型( LAB MIAs ) 。 在这种类别中, 只需对手了解预测的样本标签标签标签标签标签。 在 DAB 模型上, 对无法再培训的 DNA 模型上执行LAB MIA 的对手进行防御, 仍然是一个未解决的问题。 我们展示了LLLL, LAB 的轻重防御。 LLLLL 工作, 模型的模型决定没有改变。 与 我们的样本相比, IMA 和 LDA 的运行率, 显示我们运行的RA 成功率 。