Machine unlearning aims to erase the impact of specific training samples upon deleted requests from a trained model. Re-training the model on the retained data after deletion is an effective but not efficient way due to the huge number of model parameters and re-training samples. To speed up, a natural way is to reduce such parameters and samples. However, such a strategy typically leads to a loss in model performance, which poses the challenge that increasing the unlearning efficiency while maintaining acceptable performance. In this paper, we present a novel network, namely \textit{LegoNet}, which adopts the framework of ``fixed encoder + multiple adapters''. We fix the encoder~(\ie the backbone for representation learning) of LegoNet to reduce the parameters that need to be re-trained during unlearning. Since the encoder occupies a major part of the model parameters, the unlearning efficiency is significantly improved. However, fixing the encoder empirically leads to a significant performance drop. To compensate for the performance loss, we adopt the ensemble of multiple adapters, which are independent sub-models adopted to infer the prediction by the encoding~(\ie the output of the encoder). Furthermore, we design an activation mechanism for the adapters to further trade off unlearning efficiency against model performance. This mechanism guarantees that each sample can only impact very few adapters, so during unlearning, parameters and samples that need to be re-trained are both reduced. The empirical experiments verify that LegoNet accomplishes fast and exact unlearning while maintaining acceptable performance, synthetically outperforming unlearning baselines.
翻译:机器不学习的目的是在经过培训的模型被删除后消除特定培训样本的影响。 在删除后对保留数据模型进行再培训是一种有效但非高效的方法,因为模型参数和再培训样本数量巨大。 要加快速度,自然的方法是减少此类参数和样本。 然而,这样的战略通常会导致模型性能的丧失,这带来了提高不学习效率同时又保持可接受的性能的挑战。在本文件中,我们提出了一个新颖的网络,即\ textit{LegoNet},它采用了“固定编码+多个调试器”的框架。我们修补了勒高网络的快速编码 ~(这是代表学习的骨干), 以减少在未学习期间需要再培训的参数。由于编码占据了模型参数的主要部分,因此不学习效率得到显著提高。然而,在修正编码的精细化性能导致显著的性能下降。为了补偿性能损失,我们采用了多个调试样的组合,这是独立的分级缩缩缩图,我们要根据不断更新的性能机制来进行测试。