Partial Reasoning in Language Models: Search and Refinement Guided by Uncertainty

📅 2026-01-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models often suffer from insufficient reliability in multi-step reasoning tasks due to error accumulation. This work proposes a dynamic intervention mechanism based on the entropy of the output distribution: when entropy during generation becomes excessively high—indicating high uncertainty—the autoregressive process is temporarily halted, and a local search is performed in the latent space over the current reasoning path, followed by refinement via Soft Reasoning. For the first time, output entropy is leveraged as a signal to trigger localized optimization. The approach achieves or surpasses state-of-the-art Soft Reasoning performance on four benchmarks—GSM8K, GSM-Hard, SVAMP, and StrategyQA—demonstrating the effectiveness of entropy-guided, selective reasoning refinement.

Technology Category

Application Category

📝 Abstract
The use of Large Language Models (LLMs) for reasoning and planning tasks has drawn increasing attention in Artificial Intelligence research. Despite their remarkable progress, these models still exhibit limitations in multi-step inference scenarios, particularly in mathematical and logical reasoning. We introduce PREGU (Partial Reasoning Guided by Uncertainty). PREGU monitors the entropy of the output distribution during autoregressive generation and halts the process whenever entropy exceeds a defined threshold, signaling uncertainty. From that point, a localized search is performed in the latent space to refine the partial reasoning and select the most coherent answer, using the Soft Reasoning method. Experiments conducted with LLaMA-3-8B, Mistral-7B, and Qwen2-7B across four reasoning benchmarks (GSM8K, GSM-Hard, SVAMP, and StrategyQA) showed performance greater than or similar to Soft Reasoning, indicating that entropy can serve as an effective signal to trigger selective refinement during reasoning.
Problem

Research questions and friction points this paper is trying to address.

Large Language Models
multi-step reasoning
uncertainty
mathematical reasoning
logical reasoning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uncertainty-guided reasoning
Entropy-based halting
Partial reasoning
Latent space search
Soft Reasoning
🔎 Similar Papers
No similar papers found.
M
Murilo L. da Luz
Advanced Knowledge Center for Immersive Technologies (AKCIT); Federal University of Goiás, Brazil
B
Bruno Brandao
Advanced Knowledge Center for Immersive Technologies (AKCIT); Federal University of Goiás, Brazil
Luana Martins
Luana Martins
Postdoctoral Researcher, University of Salerno (UNISA)
Software Engineering
G
Gustavo Oliveira
Advanced Knowledge Center for Immersive Technologies (AKCIT); Federal University of Goiás, Brazil
B
Bryan de Oliveira
Advanced Knowledge Center for Immersive Technologies (AKCIT); Federal University of Goiás, Brazil
L
L. Melo
OATML, University of Oxford; Advanced Knowledge Center for Immersive Technologies (AKCIT); Federal University of Goiás, Brazil
T
Telma Soares
Advanced Knowledge Center for Immersive Technologies (AKCIT); Federal University of Goiás, Brazil