In recent years, unlearning techniques, which are methods for inducing a model to "forget" previously learned information, have attracted attention as a way to address privacy and copyright concerns in large language models (LLMs) and large multimodal models (LMMs). While several unlearning benchmarks have been established for LLMs, a practical evaluation framework for unlearning in LMMs has been less explored. Specifically, existing unlearning benchmark for LMMs considers only scenarios in which the model is required to unlearn fine-tuned knowledge through a single unlearning operation. In this study, we introduce PULSE protocol for realistic unlearning scenarios for LMMs by introducing two critical perspectives: (i) Pre-trained knowledge Unlearning for analyzing the effect across different knowledge acquisition phases and (ii) Long-term Sustainability Evaluation to address sequential requests. We then evaluate existing unlearning methods along these dimensions. Our results reveal that, although some techniques can successfully unlearn knowledge acquired through fine-tuning, they struggle to eliminate information learned during pre-training. Moreover, methods that effectively unlearn a batch of target data in a single operation exhibit substantial performance degradation when the same data are split and unlearned sequentially.
翻译:近年来,遗忘技术——即诱导模型“遗忘”先前学习信息的方法——作为解决大型语言模型(LLM)和大型多模态模型(LMM)中隐私与版权问题的一种途径,已引起广泛关注。尽管针对LLM已建立了若干遗忘基准,但针对LMM遗忘的实用评估框架却鲜有探索。具体而言,现有的LMM遗忘基准仅考虑模型通过单次遗忘操作来遗忘微调知识的场景。在本研究中,我们通过引入两个关键视角,提出了适用于LMM现实遗忘场景的PULSE协议:(i)预训练知识遗忘,用于分析不同知识获取阶段的影响;(ii)长期可持续性评估,以应对序列化遗忘请求。随后,我们沿这两个维度评估了现有的遗忘方法。我们的结果表明,尽管某些技术能成功遗忘通过微调获得的知识,却难以消除预训练阶段学习的信息。此外,那些能有效单次批量遗忘目标数据的方法,在相同数据被分割并序列化遗忘时,会表现出显著的性能下降。