Edit Less, Achieve More: Dynamic Sparse Neuron Masking for Lifelong Knowledge Editing in LLMs

📅 2025-10-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing continual knowledge editing methods for large language models suffer from error accumulation due to parameter interference, degrading both editing accuracy and generalization. This paper proposes a fine-grained neuron localization framework coupled with an entropy-guided dynamic sparse masking mechanism. First, neurons are functionally attributed and categorized into *knowledge-general* and *knowledge-specific* types. Then, only critical knowledge-specific neurons undergo sparse, adaptive weight updates—minimizing parameter perturbation while preserving model integrity. The method requires no full model retraining and maintains high editing success rates (+12.7%) and strong generalization stability (38.5% reduction in forgetting) over thousands of sequential edits—outperforming state-of-the-art approaches. Its core innovations lie in (i) interpretable, function-based neuron partitioning and (ii) an information-theoretic, sparsity-aware editing strategy that balances fidelity and plasticity.

Technology Category

Application Category

📝 Abstract
Lifelong knowledge editing enables continuous, precise updates to outdated knowledge in large language models (LLMs) without computationally expensive full retraining. However, existing methods often accumulate errors throughout the editing process, causing a gradual decline in both editing accuracy and generalization. To tackle this problem, we propose Neuron-Specific Masked Knowledge Editing (NMKE), a novel fine-grained editing framework that combines neuron-level attribution with dynamic sparse masking. Leveraging neuron functional attribution, we identify two key types of knowledge neurons, with knowledge-general neurons activating consistently across prompts and knowledge-specific neurons activating to specific prompts. NMKE further introduces an entropy-guided dynamic sparse mask, locating relevant neurons to the target knowledge. This strategy enables precise neuron-level knowledge editing with fewer parameter modifications. Experimental results from thousands of sequential edits demonstrate that NMKE outperforms existing methods in maintaining high editing success rates and preserving model general capabilities in lifelong editing.
Problem

Research questions and friction points this paper is trying to address.

Enables continuous knowledge updates in LLMs without full retraining
Addresses error accumulation during sequential editing processes
Maintains editing accuracy while preserving model generalization capabilities
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic sparse neuron masking for precise editing
Identifies knowledge-general and knowledge-specific neurons
Entropy-guided masking reduces parameter modification needs
🔎 Similar Papers
No similar papers found.
Jinzhe Liu
Jinzhe Liu
Ph.D. student, Institute of Computing Technology, Chinese Academy of Sciences
MultimodalLarge language modelKnowledge editing
J
Junshu Sun
Key Lab of Intell. Info. Process., Inst. of Comput. Tech., CAS
S
Shufan Shen
Key Lab of Intell. Info. Process., Inst. of Comput. Tech., CAS
C
Chenxue Yang
Agriculture Information Institute, CAAS
S
Shuhui Wang
Key Lab of Intell. Info. Process., Inst. of Comput. Tech., CAS