Machine learning models trained on private datasets have been shown to leak their private data. While recent work has found that the average data point is rarely leaked, the outlier samples are frequently subject to memorization and, consequently, privacy leakage. We demonstrate and analyse an Onion Effect of memorization: removing the "layer" of outlier points that are most vulnerable to a privacy attack exposes a new layer of previously-safe points to the same attack. We perform several experiments to study this effect, and understand why it occurs. The existence of this effect has various consequences. For example, it suggests that proposals to defend against memorization without training with rigorous privacy guarantees are unlikely to be effective. Further, it suggests that privacy-enhancing technologies such as machine unlearning could actually harm the privacy of other users.
翻译:在私人数据集方面受过培训的机器学习模型被证明泄露了私人数据。虽然最近的工作发现,平均数据点很少泄漏,但外部样本经常被记忆化,因此隐私泄漏。我们演示和分析了记忆化的Onion效应:消除最容易受到隐私攻击的“层”外点,暴露出新的一层先前安全点。我们进行了一些实验,以研究这一效应,并了解其为何发生。这种效应的存在产生了各种后果。例如,它表明,在没有严格隐私保障培训的情况下防范记忆化的建议不太可能有效。此外,它还表明,机器不学习等增强隐私的技术实际上可能损害其他用户的隐私。