Machine unlearning refers to the task of removing a subset of training data, thereby removing its contributions to a trained model. Approximate unlearning are one class of methods for this task which avoid the need to retrain the model from scratch on the retained data. Bayes' rule can be used to cast approximate unlearning as an inference problem where the objective is to obtain the updated posterior by dividing out the likelihood of deleted data. However this has its own set of challenges as one often doesn't have access to the exact posterior of the model parameters. In this work we examine the use of the Laplace approximation and Variational Inference to obtain the updated posterior. With a neural network trained for a regression task as the guiding example, we draw insights on the applicability of Bayesian unlearning in practical scenarios.
翻译:机器不学习是指删除一组培训数据的任务,从而消除其对经过培训的模式的贡献。近似不学习是这项任务的一种方法,避免了将模型从零开始从保留的数据上重新培训。Bayes的规则可以用来将近似不学习作为一个推论问题,因为目标是通过分离被删除数据的可能性来获得更新的后继数据。然而,这本身也有一系列挑战,因为人们往往无法获得模型参数的精确后继参数。在这项工作中,我们研究了使用拉普尔近似和变推论来获取更新的后继数据。用一个受过回归任务训练的神经网络作为指导,我们深入了解了巴耶斯语不学习在实际情景中的适用性。