We study the problem of machine unlearning and identify a notion of algorithmic stability, Total Variation (TV) stability, which we argue, is suitable for the goal of exact unlearning. For convex risk minimization problems, we design TV-stable algorithms based on noisy Stochastic Gradient Descent (SGD). Our key contribution is the design of corresponding efficient unlearning algorithms, which are based on constructing a (maximal) coupling of Markov chains for the noisy SGD procedure. To understand the trade-offs between accuracy and unlearning efficiency, we give upper and lower bounds on excess empirical and populations risk of TV stable algorithms for convex risk minimization. Our techniques generalize to arbitrary non-convex functions, and our algorithms are differentially private as well.
翻译:我们研究机器不学习的问题,并找出算法稳定性、完全变异(TV)稳定性的概念,我们认为,这适合于精确不学习的目标。对于曲线风险最小化问题,我们设计基于噪音的电磁渐变源(SGD)的电视稳定算法。我们的主要贡献是设计相应的高效不学习算法,这些算法基于为噪音的 SGD 程序构建马可夫链的(最大)联动。为了理解准确性与不学习效率之间的权衡,我们给电视稳定算法过多的经验和人口风险设定了上限和下限,以尽量减少曲线风险。我们的技术一般适用于任意的非电流功能,我们的算法也有差异性私密性。