Deep machine unlearning is the problem of removing the influence of a cohort of data from the weights of a trained deep model. This challenge is enjoying increasing attention due to the widespread use of neural networks in applications involving user data: allowing users to exercise their `right to be forgotten' necessitates an effective unlearning algorithm. However, deleting data from models is also of interest in practice for other applications where individual user privacy is not necessarily a consideration: removing biases, out-of-date examples, outliers, or noisy labels, and different such applications come with different desiderata. We propose a new unlearning algorithm (coined SCRUB) and conduct a comprehensive experimental evaluation against several previous state-of-the-art models. The results reveal that SCRUB is consistently a top performer across three different metrics for measuring unlearning quality, reflecting different application scenarios, while not degrading the model's performance.
翻译:深层机器不学习是将一组数据的影响从经过训练的深层模型的重量中去除的问题。由于在涉及用户数据的应用程序中广泛使用神经网络,这一挑战正日益受到关注:允许用户行使其“被遗忘的权利”需要一种有效的不学习算法。然而,在个人用户隐私不一定考虑的其他应用程序中,从模型中删除数据也具有实际意义:消除偏见、过时的范例、外部标签或吵闹的标签,以及不同应用程序出现不同的偏差。我们提出了一个新的未学习算法(coined SCRUB),并对以前的几个最先进的模型进行全面的实验性评估。结果显示,SCRUB始终是衡量不学习质量、反映不同应用情景的三种不同标准中最优秀的,同时不贬低模型的性能。