🤖 AI Summary
Large reasoning models (LRMs) suffer from token redundancy and “overthinking” errors due to reliance on lengthy chain-of-thought (CoT) prompting. Method: We propose a novel paradigm integrating large language models (LLMs) with symbolic solvers, leveraging LLMs’ code-generation capability to translate reasoning tasks into executable declarative programs; structured in-context examples guide the LLM to invoke symbolic solvers for precise, constraint-aware solving. Contribution/Results: This integration proves especially effective on constraint satisfaction problems (CSPs) with limited implicit reasoning but large search spaces and frequent backtracking—e.g., the Zebra Puzzle. Under declarative prompting, CodeLlama-13B surpasses GPT-4o in accuracy. Our work is the first systematic study to characterize the efficacy boundaries of symbolic solver augmentation for LLM-based reasoning, demonstrating that lightweight models can achieve high-precision symbolic reasoning without scaling up parameters—offering a viable path toward efficient, verifiable AI reasoning.
📝 Abstract
Large Reasoning Models (LRMs) achieve strong performance on complex reasoning tasks by generating long Chains of Thought (CoTs). However, this paradigm might incur substantial token overhead, especially when models"overthink"by producing lengthy reasoning chains, which can even lead to incorrect answers. A promising direction is the symbolic-solver-integrated approach, which leverages the code generation capabilities of LLMs to translate reasoning tasks into executable code and then solve them with a symbolic solver. In this paper, we explore an open question of when the conventional long-CoT can be enhanced by symbolic solvers. Our experimental results show that the symbolic-solver-integrated method only helps when the problem requires limited implicit reasoning but involves an ample search space. The latest LLMs, like GPT-4o, show better performance on deductive problems with shallow reasoning depth, while the symbolic-solver-integrated method significantly improves the LLMs'performance in constraint satisfaction problems that require repeated backtracks. When a declarative exemplar is provided, even CodeLlama-13B can outperform GPT-4o in difficult Zebra puzzles.