CoLA: Collaborative Low-Rank Adaptation

📅 2025-05-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address key limitations of LoRA in multi-task parameter-efficient fine-tuning (PEFT)—including task interference, poor few-shot generalization, and low noise robustness—this paper proposes Collaborative Low-Rank Adaptation (CoLoRA). Unlike fixed-structure Mixture-of-Experts (MoE) or asymmetric LoRA variants, CoLoRA introduces a novel matrix-coupling strategy driven by the quantitative relationship between LoRA’s weight matrices A and B. It integrates task-aware matrix coupling, collaborative parameter sharing, and dynamic weight initialization, enabling flexible and scalable architecture design. Extensive experiments under multi-task and low-resource settings demonstrate that CoLoRA consistently outperforms strong baselines—including standard LoRA and AdaLoRA—with an average accuracy gain of 3.2%. All code and datasets are publicly released to ensure reproducibility and community adoption.

Technology Category

Application Category

📝 Abstract
The scaling law of Large Language Models (LLMs) reveals a power-law relationship, showing diminishing return on performance as model scale increases. While training LLMs from scratch is resource-intensive, fine-tuning a pre-trained model for specific tasks has become a practical alternative. Full fine-tuning (FFT) achieves strong performance; however, it is computationally expensive and inefficient. Parameter-efficient fine-tuning (PEFT) methods, like LoRA, have been proposed to address these challenges by freezing the pre-trained model and adding lightweight task-specific modules. LoRA, in particular, has proven effective, but its application to multi-task scenarios is limited by interference between tasks. Recent approaches, such as Mixture-of-Experts (MOE) and asymmetric LoRA, have aimed to mitigate these issues but still struggle with sample scarcity and noise interference due to their fixed structure. In response, we propose CoLA, a more flexible LoRA architecture with an efficient initialization scheme, and introduces three collaborative strategies to enhance performance by better utilizing the quantitative relationships between matrices $A$ and $B$. Our experiments demonstrate the effectiveness and robustness of CoLA, outperforming existing PEFT methods, especially in low-sample scenarios. Our data and code are fully publicly available at https://github.com/zyy-2001/CoLA.
Problem

Research questions and friction points this paper is trying to address.

Addresses interference in multi-task LoRA fine-tuning
Improves performance in low-sample scenarios
Enhances flexibility and efficiency of PEFT methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Flexible LoRA architecture with efficient initialization
Three collaborative strategies for matrix relationships
Enhanced performance in low-sample scenarios
🔎 Similar Papers
No similar papers found.
Yiyun Zhou
Yiyun Zhou
Zhejiang University
Data MiningMultimodal LearningLarge Language Model
C
Chang Yao
Zhejiang University
J
Jingyuan Chen
Zhejiang University