We show that the influence of a subset of the training samples can be removed -- or "forgotten" -- from the weights of a network trained on large-scale image classification tasks, and we provide strong computable bounds on the amount of remaining information after forgetting. Inspired by real-world applications of forgetting techniques, we introduce a novel notion of forgetting in mixed-privacy setting, where we know that a "core" subset of the training samples does not need to be forgotten. While this variation of the problem is conceptually simple, we show that working in this setting significantly improves the accuracy and guarantees of forgetting methods applied to vision classification tasks. Moreover, our method allows efficient removal of all information contained in non-core data by simply setting to zero a subset of the weights with minimal loss in performance. We achieve these results by replacing a standard deep network with a suitable linear approximation. With opportune changes to the network architecture and training procedure, we show that such linear approximation achieves comparable performance to the original network and that the forgetting problem becomes quadratic and can be solved efficiently even for large models. Unlike previous forgetting methods on deep networks, ours can achieve close to the state-of-the-art accuracy on large scale vision tasks. In particular, we show that our method allows forgetting without having to trade off the model accuracy.
翻译:我们显示,一组培训样本的影响可以从受过大规模图像分类任务培训的网络的重量中去除 -- -- 或“被遗忘”的影响,而我们则在忘记后对剩余信息的数量提供很强的可比较界限。在现实世界应用遗忘技术的启发下,我们引入了一种新的概念,即在混合隐私环境中遗忘,我们知道网络结构和培训程序的“核心”子并不需要被遗忘。虽然这一问题的这种差异在概念上是简单的,但我们表明,在这种环境中工作极大地提高了遗忘用于愿景分类任务的方法的准确性和保障。此外,我们的方法允许通过简单地设定零分部分重量而使业绩损失最小的分数来有效删除非核心数据中的所有信息。我们通过以适当的线性近似方式取代标准的深度网络,从而实现这些结果。随着网络结构和培训程序的适时变化,我们表明,这种线性近率取得了与原始网络相似的性能,而忘记的问题变得四分化,甚至可以有效地解决大型模型。我们的方法与以前遗忘的深层次网络方法不同,我们的方法可以使贸易的精确度接近于我们无法显示。