Large pre-trained generative models are known to occasionally provide samples that may be undesirable for various reasons. The standard way to mitigate this is to re-train the models differently. In this work, we take a different, more compute-friendly approach and investigate how to post-edit a model after training so that it forgets certain kinds of samples. We provide three different algorithms for GANs that differ on how the samples to be forgotten are described. Extensive evaluations on real-world image datasets show that our algorithms are capable of forgetting data while retaining high generation quality at a fraction of the cost of full re-training.
翻译:据了解,大型的经过培训的基因变异模型有时会提供由于各种原因可能不可取的样本。标准的缓解方法就是对模型进行不同的再培训。在这项工作中,我们采取了不同的、更便于计算的方法,并调查如何在培训后编辑模型,以便它忘记某些类型的样本。我们为GAN提供了三种不同的算法,这些算法在如何描述被遗忘的样本方面有所不同。对现实世界图像数据集的广泛评价表明,我们的算法能够忘记数据,同时保留高产质量,而完全再培训成本的一小部分。