The usage of deep learning is being escalated in many applications. Due to its outstanding performance, it is being used in a variety of security and privacy-sensitive areas in addition to conventional applications. One of the key aspects of deep learning efficacy is to have abundant data. This trait leads to the usage of data which can be highly sensitive and private, which in turn causes wariness with regard to deep learning in the general public. Membership inference attacks are considered lethal as they can be used to figure out whether a piece of data belongs to the training dataset or not. This can be problematic with regards to leakage of training data information and its characteristics. To highlight the significance of these types of attacks, we propose an enhanced methodology for membership inference attacks based on adversarial robustness, by adjusting the directions of adversarial perturbations through label smoothing under a white-box setting. We evaluate our proposed method on three datasets: Fashion-MNIST, CIFAR-10, and CIFAR-100. Our experimental results reveal that the performance of our method surpasses that of the existing adversarial robustness-based method when attacking normally trained models. Additionally, through comparing our technique with the state-of-the-art metric-based membership inference methods, our proposed method also shows better performance when attacking adversarially trained models. The code for reproducing the results of this work is available at \url{https://github.com/plll4zzx/Evaluating-Membership-Inference-Through-Adversarial-Robustness}.
翻译:深层次学习的使用在许多应用中不断升级。由于它的杰出表现,它被应用于各种安全和隐私敏感领域,除了常规应用外,还被用于各种安全和隐私敏感领域。深层次学习功效的一个关键方面是拥有丰富的数据。这一特征导致使用高度敏感和隐私的数据,这反过来又导致公众对深层次学习产生戒备。会员推断攻击被认为是致命的,因为可以用来查明某一数据是否属于培训数据集。这在培训数据信息及其特征的渗漏方面可能存在问题。为了突出这些类型的攻击的重要性,我们提出一种基于对抗性强力的强化成员推断攻击方法。我们通过在白箱设置下打平滑标签来调整对抗性干扰的方向。我们评估了我们关于三个数据集的拟议方法:法西文-MNIST、CIFAR-10和CIFAR-100。我们的实验结果表明,我们的方法在攻击通常训练的模型时,其表现超过了现有的以对抗性强度为基础的方法。此外,通过将我们经过培训的模范式模型的模化方法与现在的模范式模型相比较。