When Do Symbolic Solvers Enhance Reasoning in Large Language Models?

📅 2025-12-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large reasoning models (LRMs) suffer from token redundancy and “overthinking” errors due to reliance on lengthy chain-of-thought (CoT) prompting. Method: We propose a novel paradigm integrating large language models (LLMs) with symbolic solvers, leveraging LLMs’ code-generation capability to translate reasoning tasks into executable declarative programs; structured in-context examples guide the LLM to invoke symbolic solvers for precise, constraint-aware solving. Contribution/Results: This integration proves especially effective on constraint satisfaction problems (CSPs) with limited implicit reasoning but large search spaces and frequent backtracking—e.g., the Zebra Puzzle. Under declarative prompting, CodeLlama-13B surpasses GPT-4o in accuracy. Our work is the first systematic study to characterize the efficacy boundaries of symbolic solver augmentation for LLM-based reasoning, demonstrating that lightweight models can achieve high-precision symbolic reasoning without scaling up parameters—offering a viable path toward efficient, verifiable AI reasoning.

Technology Category

Application Category

📝 Abstract
Large Reasoning Models (LRMs) achieve strong performance on complex reasoning tasks by generating long Chains of Thought (CoTs). However, this paradigm might incur substantial token overhead, especially when models"overthink"by producing lengthy reasoning chains, which can even lead to incorrect answers. A promising direction is the symbolic-solver-integrated approach, which leverages the code generation capabilities of LLMs to translate reasoning tasks into executable code and then solve them with a symbolic solver. In this paper, we explore an open question of when the conventional long-CoT can be enhanced by symbolic solvers. Our experimental results show that the symbolic-solver-integrated method only helps when the problem requires limited implicit reasoning but involves an ample search space. The latest LLMs, like GPT-4o, show better performance on deductive problems with shallow reasoning depth, while the symbolic-solver-integrated method significantly improves the LLMs'performance in constraint satisfaction problems that require repeated backtracks. When a declarative exemplar is provided, even CodeLlama-13B can outperform GPT-4o in difficult Zebra puzzles.
Problem

Research questions and friction points this paper is trying to address.

Enhancing reasoning efficiency in Large Language Models
Determining when symbolic solvers outperform long reasoning chains
Improving performance on constraint satisfaction and search problems
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLMs generate code for symbolic solver execution
Symbolic solvers enhance constraint satisfaction with backtracking
Method improves limited implicit reasoning with large search space
🔎 Similar Papers
No similar papers found.