Towards Benign Memory Forgetting for Selective Multimodal Large Language Model Unlearning

πŸ“… 2025-11-25
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing privacy forgetting methods for multimodal large language models (MLLMs) often degrade general visual understanding capabilities. Method: We propose a selective forgetting framework featuring the Sculpted Memory Forgetting Adapter, which integrates an anchor-guided masking mechanism and memory isolation to precisely erase sensitive knowledge; we further enhance forgetting robustness via rejection-response training. Contribution/Results: We introduce S-MLLMUn Benchβ€”the first benchmark jointly evaluating forgetting efficacy and visual understanding preservation. Experiments across multiple MLLMs demonstrate 100% sensitive information removal while retaining over 98% of image understanding performance, significantly outperforming state-of-the-art forgetting approaches.

Technology Category

Application Category

πŸ“ Abstract
Multimodal Large Language Models (MLLMs) achieve remarkable capabilities but can inadvertently memorize privacy-sensitive information. Although existing unlearning methods can remove such knowledge, they fail to achieve benign forgetting because they often degrade the model's general image understanding performance. To address this, we propose the Sculpted Memory Forgetting Adapter (SMFA), which confines forgetting to targeted memory regions while preserving overall capabilities. SMFA first fine-tunes the model to replace sensitive responses with refusals, yielding a memory forgetting adapter, and then applies a retaining anchor-guided masking mechanism to prevent interference with unrelated knowledge and understanding ability. To systematically evaluate selective MLLM unlearning, we introduce S-MLLMUn Bench, the first benchmark designed to jointly assess the removal of sensitive knowledge and retention of general visual understanding. Extensive experiments show that, unlike prior methods, SMFA achieves precise and controllable unlearning while maintaining the model's foundational image understanding.
Problem

Research questions and friction points this paper is trying to address.

Preventing privacy leakage from memorized sensitive data
Maintaining general image understanding during unlearning
Achieving precise targeted forgetting without performance degradation
Innovation

Methods, ideas, or system contributions that make the work stand out.

SMFA confines forgetting to targeted memory regions
SMFA uses retaining anchor-guided masking mechanism
SMFA achieves precise unlearning while maintaining understanding
πŸ”Ž Similar Papers
No similar papers found.
Z
Zhen Zeng
Hefei University of Technology
L
Leijiang Gu
Hefei University of Technology
Z
Zhangling Duan
Hefei Comprehensive National Science Center
F
Feng Li
Hefei University of Technology
Zenglin Shi
Zenglin Shi
Professor of Artificial Intelligence, Hefei University of Technology
Deep LearningComputer VisionMachine LearningMultimedia
Cees G. M. Snoek
Cees G. M. Snoek
Professor of Computer Science, University of Amsterdam
Video Understanding:computer visionmultimodal learningmachine learningartificial intelligence
M
Meng Wang
Hefei University of Technology