FRoD: Full-Rank Efficient Fine-Tuning with Rotational Degrees for Fast Convergence

📅 2025-12-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing parameter-efficient fine-tuning (PEFT) methods—such as LoRA—are constrained by low-rank approximations, leading to slow convergence, limited expressivity, and difficulty in capturing complex downstream task patterns. To address this, we propose HR-LoRA, a novel PEFT framework integrating hierarchical joint decomposition with rotational degree-of-freedom modeling. Its core innovation lies in a globally shared low-dimensional basis matrix coupled with task-adaptive sparse scaling perturbations, enabling near-full-rank parameter updates while maintaining an extremely low number of trainable parameters. This design bridges the expressivity gap between PEFT and full fine-tuning. Evaluated across 20 cross-modal benchmarks, HR-LoRA achieves performance on par with full fine-tuning using only 1.72% trainable parameters, while substantially accelerating convergence and improving out-of-distribution robustness.

Technology Category

Application Category

📝 Abstract
Parameter-efficient fine-tuning (PEFT) methods have emerged as a practical solution for adapting large foundation models to downstream tasks, reducing computational and memory costs by updating only a small subset of parameters. Among them, approaches like LoRA aim to strike a balance between efficiency and expressiveness, but often suffer from slow convergence and limited adaptation capacity due to their inherent low-rank constraints. This trade-off hampers the ability of PEFT methods to capture complex patterns needed for diverse tasks. To address these challenges, we propose FRoD, a novel fine-tuning method that combines hierarchical joint decomposition with rotational degrees of freedom. By extracting a globally shared basis across layers and injecting sparse, learnable perturbations into scaling factors for flexible full-rank updates, FRoD enhances expressiveness and efficiency, leading to faster and more robust convergence. On 20 benchmarks spanning vision, reasoning, and language understanding, FRoD matches full model fine-tuning in accuracy, while using only 1.72% of trainable parameters under identical training budgets.
Problem

Research questions and friction points this paper is trying to address.

Addresses slow convergence in parameter-efficient fine-tuning methods
Overcomes limited adaptation capacity due to low-rank constraints
Enhances expressiveness and efficiency for diverse tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hierarchical joint decomposition with rotational degrees
Sparse learnable perturbations in scaling factors
Full-rank updates for faster robust convergence
🔎 Similar Papers
No similar papers found.
G
Guoan Wan
School of Computer Science and Engineering, Beihang University, China
T
Tianyu Chen
School of Computer Science and Engineering, Beihang University, China
F
Fangzheng Feng
School of Electronic Information and Communications, Huazhong University of Science and Technoledgy, China
Haoyi Zhou
Haoyi Zhou
Associate Professor, Beihang University
Machine LearningData MiningTime-series
Runhua Xu
Runhua Xu
Beihang University | former RSM@IBM Research
privacy-enhancing tech.security/privacy in AI/MLapplied cryptoblockchain