The vulnerability of machine learning models to membership inference attacks has received much attention in recent years. However, existing attacks mostly remain impractical due to having high false positive rates, where non-member samples are often erroneously predicted as members. This type of error makes the predicted membership signal unreliable, especially since most samples are non-members in real world applications. In this work, we argue that membership inference attacks can benefit drastically from \emph{difficulty calibration}, where an attack's predicted membership score is adjusted to the difficulty of correctly classifying the target sample. We show that difficulty calibration can significantly reduce the false positive rate of a variety of existing attacks without a loss in accuracy.
翻译:近年来,机器学习模式容易成为成员推论攻击的弱点引起了人们的极大关注,然而,现有的攻击大多仍然是不切实际的,因为假正率很高,非成员样本往往被错误地预测为成员。这类错误使得预测会籍信号不可靠,特别是因为大多数样本是真实世界应用程序中的非成员。 在这项工作中,我们争辩说,会籍推断攻击可以从\emph{困难校准}中得到极大的好处,因为攻击的预测会籍得分调整到对目标样本进行正确分类的难度。我们表明,难以校准可以大大降低现有各种攻击的假正率,而不会失去准确性。