Structure-Learnable Adapter Fine-Tuning for Parameter-Efficient Large Language Models

📅 2025-09-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address parameter redundancy, structural rigidity, and insufficient task adaptability in large language model (LLM) fine-tuning, this paper proposes a learnable-structure adapter tuning method. Under frozen backbone parameters, it jointly optimizes adapter insertion positions, activation paths, and module compositions via differentiable gating and structural sparsity control, dynamically constructing task-specific, efficient substructures. Its key innovation lies in formulating architecture search as a differentiable optimization problem, integrating sensitivity analysis to quantify the impact of sparsity, noise, and data perturbations. Experiments across multiple natural language understanding benchmarks demonstrate that the method significantly outperforms mainstream parameter-efficient fine-tuning approaches—achieving higher accuracy, superior parameter compression ratios, and enhanced robustness against input noise.

Technology Category

Application Category

📝 Abstract
This paper addresses the issues of parameter redundancy, rigid structure, and limited task adaptability in the fine-tuning of large language models. It proposes an adapter-based fine-tuning method built on a structure-learnable mechanism. By introducing differentiable gating functions and structural sparsity control variables, the method enables automatic optimization of adapter insertion points, activation paths, and module combinations. This allows the model to adjust its structure flexibly in multi-task settings to match different task characteristics. With the backbone parameters kept frozen, the method uses a structure search mechanism to guide the dynamic construction of task-specific efficient substructures during training. This significantly improves parameter utilization and representational capacity. In addition, the paper designs a set of sensitivity analysis experiments to systematically evaluate the effects of sparsity weight, noise injection ratio, and data perturbation on model performance. These experiments verify the stability and robustness of the proposed method across various multi-task natural language understanding tasks. The experimental results show that the proposed method outperforms mainstream parameter-efficient tuning techniques on multiple tasks. It achieves a better balance among accuracy, compression rate, and robustness to noise and perturbation.
Problem

Research questions and friction points this paper is trying to address.

Reduces parameter redundancy in fine-tuning large models
Enhances task adaptability through learnable adapter structures
Improves robustness against noise and data perturbations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Structure-learnable adapter fine-tuning mechanism
Differentiable gating functions with sparsity control
Dynamic construction of task-specific efficient substructures
🔎 Similar Papers
Ming Gong
Ming Gong
Key laboratory of quantum information, USTC
quantum informationquantum dottopological quantum phase transitionultracold atomsFFLO
Y
Yingnan Deng
Georgia Institute of Technology Atlanta, USA
N
Nia Qi
Independent Author Pittsburgh, USA
Yujun Zou
Yujun Zou
University of California, Berkeley
Large Language ModelsAI Agent
Z
Zhihao Xue
Y
Yun Zi
Georgia Institute of Technology Atlanta, USA