Merlin's Whisper: Enabling Efficient Reasoning in LLMs via Black-box Adversarial Prompting

📅 2025-10-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large reasoning models (LRMs) exhibit strong reasoning capabilities but suffer from excessive computational overhead and latency due to lengthy chain-of-thought (CoT) generation—a phenomenon termed “overthinking.” This work proposes AdvPrompt, the first black-box adversarial prompting framework specifically designed to mitigate overthinking in LRMs. AdvPrompt compresses reasoning paths via multi-perspective input perturbation and iterative prompt refinement, without requiring model access, gradient information, or architectural modifications—ensuring compatibility with both open- and closed-weight models across scales and architectures. Experiments on GSM8K and MATH-500 demonstrate that AdvPrompt reduces average token consumption by ~40% while preserving accuracy. Notably, Qwen3’s response length on GSM8K easy questions is reduced by 3×; Claude-3.7 and Gemini-2.5 achieve 35% and 47% token savings, respectively, on MATH-500.

Technology Category

Application Category

📝 Abstract
Large reasoning models (LRMs) have demonstrated remarkable proficiency in tackling complex reasoning tasks through step-by-step thinking. However, such a lengthy reasoning process incurs substantial computational and latency overheads, hindering the practical deployment of these models. In this work, we present a new perspective on mitigating overthinking in LRMs via black-box adversarial prompting. By treating both open-source LRMs and closed-source APIs as black-box communicators, we investigate how to elicit concise responses without sacrificing accuracy. We introduce AdvPrompt, an iterative refinement framework that generates high-quality adversarial prompts from diverse perspectives. Experiments across multiple benchmarks demonstrate that AdvPrompt consistently reduces token usage while preserving performance. Notably, AdvPrompt achieves a 3x reduction in average response length on simple GSM8K questions for the Qwen3 model series, and delivers an average ~40% token reduction across four benchmarks. For closed-source APIs, AdvPrompt reduces token usage on MATH-500 by 35% for Claude-3.7 and 47% for Gemini-2.5. Further analysis reveals the generalizability of AdvPrompt across various model scales and families, underscoring the potential of black-box prompting as a practical and effective strategy for enhancing LRM efficiency.
Problem

Research questions and friction points this paper is trying to address.

Reducing computational overhead in large reasoning models
Minimizing token usage while preserving model accuracy
Addressing overthinking in black-box LLMs via adversarial prompting
Innovation

Methods, ideas, or system contributions that make the work stand out.

Black-box adversarial prompting reduces reasoning overhead
Iterative refinement framework generates efficient adversarial prompts
Token usage reduction while preserving model performance accuracy
🔎 Similar Papers
No similar papers found.
Heming Xia
Heming Xia
Natural Language Processing Group, The Hong Kong Polytechnic University
Natural Language ProcessingLarge Language Models
Cunxiao Du
Cunxiao Du
Research Scientist at Sea AI Lab
NLPLLM Inference
R
Rui Li
Peking University
C
Chak Tou Leong
Department of Computing, The Hong Kong Polytechnic University
Y
Yongqi Li
Department of Computing, The Hong Kong Polytechnic University
W
Wenjie Li
Department of Computing, The Hong Kong Polytechnic University