Shadow Unlearning: A Neuro-Semantic Approach to Fidelity-Preserving Faceless Forgetting in LLMs

๐Ÿ“… 2026-01-07
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work proposes the Shadow Unlearning paradigm to fulfill the โ€œright to be forgottenโ€ mandated by regulations such as the GDPR, enabling efficient removal of the influence of specific training data from large language models without accessing the original personally identifiable information (PII). The core method, Neuro-Semantic Projector Unlearning (NSPU), leverages neural-semantic projection to perform approximate unlearning on anonymized or synthetic forget samples, effectively balancing privacy preservation with model utility. Experimental results demonstrate that NSPU significantly outperforms existing approaches across multiple large language models, achieving stronger unlearning efficacy and enhanced privacy guarantees while preserving high-fidelity knowledge retention. Moreover, NSPU improves computational efficiency by at least an order of magnitude compared to current methods.

Technology Category

Application Category

๐Ÿ“ Abstract
Machine unlearning aims to selectively remove the influence of specific training samples to satisfy privacy regulations such as the GDPR's'Right to be Forgotten'. However, many existing methods require access to the data being removed, exposing it to membership inference attacks and potential misuse of Personally Identifiable Information (PII). We address this critical challenge by proposing Shadow Unlearning, a novel paradigm of approximate unlearning, that performs machine unlearning on anonymized forget data without exposing PII. We further propose a novel privacy-preserving framework, Neuro-Semantic Projector Unlearning (NSPU) to achieve Shadow unlearning. To evaluate our method, we compile Multi-domain Fictitious Unlearning (MuFU) forget set across five diverse domains and introduce an evaluation stack to quantify the trade-off between knowledge retention and unlearning effectiveness. Experimental results on various LLMs show that NSPU achieves superior unlearning performance, preserves model utility, and enhances user privacy. Additionally, the proposed approach is at least 10 times more computationally efficient than standard unlearning approaches. Our findings foster a new direction for privacy-aware machine unlearning that balances data protection and model fidelity.
Problem

Research questions and friction points this paper is trying to address.

machine unlearning
privacy-preserving
Right to be Forgotten
Personally Identifiable Information
large language models
Innovation

Methods, ideas, or system contributions that make the work stand out.

Shadow Unlearning
Neuro-Semantic Projector Unlearning
privacy-preserving machine unlearning
fidelity-preserving forgetting
approximate unlearning
๐Ÿ”Ž Similar Papers
No similar papers found.