The right to be forgotten has been legislated in many countries but the enforcement in machine learning would cause unbearable costs: companies may need to delete whole models trained from massive resources because of single individual requests. Existing works propose to remove the influence of the requested datums on the learned models via its influence function which is no longer naturally well-defined in Bayesian inference. To address this problem, this paper proposes a {\it Bayesian inference forgetting} (BIF) framework to extend the applicable domain to Bayesian inference. In the BIF framework, we develop forgetting algorithms for variational inference and Markov chain Monte Carlo. We show that our algorithms can provably remove the influence of single datums on the learned models. Theoretical analysis demonstrates that our algorithms have guaranteed generalizability. Experiments of Gaussian mixture models on the synthetic dataset and Bayesian neural networks on the Fashion-MNIST dataset verify the feasibility of our methods. The source code package is available at \url{https://github.com/fshp971/BIF}.
翻译:在许多国家,人们已经立法了被遗忘的权利,但机器学习的强制执行将造成无法承受的成本:公司可能需要根据单个个人的要求,删除从大量资源中培训出来的整个模型;现有工作提议,通过影响功能,消除所要求的数据对知识模型的影响,这种影响功能在巴伊西亚的推论中已不再自然地明确界定。为了解决这一问题,本文件提议了一个“BIF”框架,将适用的范围扩大到巴伊西亚的推断。在BIF框架内,我们开发了用于变异推断的算法和Markov连锁的Monte Carlo。我们表明,我们的算法可以消除单一数据对所学模型的影响。理论分析表明,我们的算法已经保证了可概括性。关于合成数据集的高斯混合模型实验和Fashion-MNIST数据集上的巴伊西亚神经网络证实了我们的方法的可行性。源代码包可在以下https://github.com/fsh71/BIF}。