Iterative Prompt Relocation for Distribution-Adaptive Visual Prompt Tuning

📅 2025-03-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the poor cross-task adaptability of visual prompt tuning (VPT) caused by fixed prompt distribution across Transformer layers, this paper proposes PRO-VPT—a distribution-adaptive, iterative prompt relocalization framework. We first formally define Adaptive Distribution Optimization (ADO), revealing its nested optimization structure inherent in VPT. PRO-VPT employs a two-stage strategy: (i) gradient-sensitivity-based identification and pruning of idle prompts, and (ii) optimal block-wise redeployment guided by inter-layer semantic alignment. A learnable relocalization module, jointly optimized with the backbone in an iterative manner, dynamically adjusts prompt allocation across Transformer layers. Evaluated on VTAB-1k, PRO-VPT achieves an average accuracy gain of 1.6% over standard VPT, establishing new state-of-the-art performance among prompt-tuning methods.

Technology Category

Application Category

📝 Abstract
Visual prompt tuning (VPT) provides an efficient and effective solution for adapting pre-trained models to various downstream tasks by incorporating learnable prompts. However, most prior art indiscriminately applies a fixed prompt distribution across different tasks, neglecting the importance of each block differing depending on the task. In this paper, we investigate adaptive distribution optimization (ADO) by addressing two key questions: (1) How to appropriately and formally define ADO, and (2) How to design an adaptive distribution strategy guided by this definition? Through in-depth analysis, we provide an affirmative answer that properly adjusting the distribution significantly improves VPT performance, and further uncover a key insight that a nested relationship exists between ADO and VPT. Based on these findings, we propose a new VPT framework, termed PRO-VPT (iterative Prompt RelOcation-based VPT), which adaptively adjusts the distribution building upon a nested optimization formulation. Specifically, we develop a prompt relocation strategy for ADO derived from this formulation, comprising two optimization steps: identifying and pruning idle prompts, followed by determining the optimal blocks for their relocation. By iteratively performing prompt relocation and VPT, our proposal adaptively learns the optimal prompt distribution, thereby unlocking the full potential of VPT. Extensive experiments demonstrate that our proposal significantly outperforms state-of-the-art VPT methods, e.g., PRO-VPT surpasses VPT by 1.6% average accuracy, leading prompt-based methods to state-of-the-art performance on the VTAB-1k benchmark. The code is available at https://github.com/ckshang/PRO-VPT.
Problem

Research questions and friction points this paper is trying to address.

Adapts pre-trained models to various tasks using learnable prompts.
Optimizes prompt distribution for improved visual prompt tuning performance.
Proposes iterative prompt relocation to enhance task-specific adaptability.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive distribution optimization for VPT
Iterative prompt relocation strategy
Nested optimization formulation enhances VPT
🔎 Similar Papers
No similar papers found.
C
Chikai Shang
Guangdong University of Technology
M
Mengke Li
Shenzhen University
Y
Yiqun Zhang
Guangdong University of Technology
Z
Zhen Chen
Yale University
Jinlin Wu
Jinlin Wu
Institute of Automation,Chinese Academy of Sciences
Fangqing Gu
Fangqing Gu
Guangdong University of Technology
Evolutionary AlgorithmMachine Learning
Y
Yang Lu
Xiamen University
Y
Yiu-ming Cheung
Hong Kong Baptist University