🤖 AI Summary
To address content-style feature entanglement arising from multi-LoRA fusion in text-to-image models, this paper proposes QR-LoRA—a parameter-efficient fine-tuning framework based on QR decomposition. It decomposes weight updates into an orthogonal matrix **Q**, which suppresses cross-attribute interference, and an upper-triangular matrix **R**, which encodes attribute-specific transformations; a learnable increment ΔR enables clean, contamination-free merging of multiple tasks. Compared to standard LoRA, QR-LoRA reduces parameter count by 50% while inherently satisfying orthogonality constraints. Experiments demonstrate substantial improvements in both content-style disentangled generation and secure multi-model fusion—achieving higher generation quality and disentanglement accuracy—without compromising efficiency or controllability.
📝 Abstract
Existing text-to-image models often rely on parameter fine-tuning techniques such as Low-Rank Adaptation (LoRA) to customize visual attributes. However, when combining multiple LoRA models for content-style fusion tasks, unstructured modifications of weight matrices often lead to undesired feature entanglement between content and style attributes. We propose QR-LoRA, a novel fine-tuning framework leveraging QR decomposition for structured parameter updates that effectively separate visual attributes. Our key insight is that the orthogonal Q matrix naturally minimizes interference between different visual features, while the upper triangular R matrix efficiently encodes attribute-specific transformations. Our approach fixes both Q and R matrices while only training an additional task-specific $ΔR$ matrix. This structured design reduces trainable parameters to half of conventional LoRA methods and supports effective merging of multiple adaptations without cross-contamination due to the strong disentanglement properties between $ΔR$ matrices. Experiments demonstrate that QR-LoRA achieves superior disentanglement in content-style fusion tasks, establishing a new paradigm for parameter-efficient, disentangled fine-tuning in generative models.