LoRA-Based Continual Learning with Constraints on Critical Parameter Changes

📅 2025-04-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address catastrophic forgetting in LoRA-based continual learning—where orthogonal fine-tuning still induces drift in critical pre-trained parameters—we propose a stability-plasticity-balanced approach. Specifically, we introduce the first explicit freezing of task-critical parameter matrices from prior tasks within Vision Transformers (ViTs), coupled with a novel QR-decomposition-based LoRA composition paradigm (LoRAC) to enhance representational capacity and task adaptability of orthogonal LoRA. Our method integrates parameter importance estimation, orthogonal constraint optimization, and LoRA fine-tuning. Evaluated on Split CIFAR-100, it achieves a 6.35% absolute accuracy gain and reduces average forgetting by 3.24%, establishing new state-of-the-art performance in continual learning.

Technology Category

Application Category

📝 Abstract
LoRA-based continual learning represents a promising avenue for leveraging pre-trained models in downstream continual learning tasks. Recent studies have shown that orthogonal LoRA tuning effectively mitigates forgetting. However, this work unveils that under orthogonal LoRA tuning, the critical parameters for pre-tasks still change notably after learning post-tasks. To address this problem, we directly propose freezing the most critical parameter matrices in the Vision Transformer (ViT) for pre-tasks before learning post-tasks. In addition, building on orthogonal LoRA tuning, we propose orthogonal LoRA composition (LoRAC) based on QR decomposition, which may further enhance the plasticity of our method. Elaborate ablation studies and extensive comparisons demonstrate the effectiveness of our proposed method. Our results indicate that our method achieves state-of-the-art (SOTA) performance on several well-known continual learning benchmarks. For instance, on the Split CIFAR-100 dataset, our method shows a 6.35% improvement in accuracy and a 3.24% reduction in forgetting compared to previous methods. Our code is available at https://github.com/learninginvision/LoRAC-IPC.
Problem

Research questions and friction points this paper is trying to address.

Mitigates forgetting in continual learning tasks
Freezes critical ViT parameters for pre-tasks
Enhances plasticity via orthogonal LoRA composition
Innovation

Methods, ideas, or system contributions that make the work stand out.

Freezing critical ViT parameter matrices
Orthogonal LoRA composition (LoRAC) via QR
Enhancing plasticity in continual learning
🔎 Similar Papers
No similar papers found.
S
Shimou Ling
University of Electronic Science and Technology of China, Chengdu, China
L
Liang Zhang
University of Electronic Science and Technology of China, Chengdu, China
Jiangwei Zhao
Jiangwei Zhao
University of Electronic Science and Technology of China
diffusioncontinual learning
Lili Pan
Lili Pan
Associate Professor, University of Electronic Science and Technology of China
Computer visionMachine learning
H
Hongliang Li
University of Electronic Science and Technology of China, Chengdu, China