🤖 AI Summary
This work investigates the generalization capability of test-time scaling (TTS) for theoretical physics reasoning and analyzes its differential applicability compared to mathematical reasoning tasks (e.g., AIME). Addressing the structured, domain-specific nature of physics problems, we propose a symbolic weak verification framework: during parallel sampling, it incorporates stepwise, physics-law-guided symbolic validation to enhance reasoning consistency and reliability. We systematically evaluate mainstream TTS methods on TPBench—a newly constructed benchmark for theoretical physics—and demonstrate that our approach significantly outperforms existing baselines. Cross-domain validation on AIME further confirms its transferability. Our core contribution is the first integration of structure-aware symbolic verification into the TTS paradigm, yielding an interpretable and robust reasoning enhancement pathway tailored to scientific reasoning tasks.
📝 Abstract
Large language models (LLMs) have shown strong capabilities in complex reasoning, and test-time scaling techniques can enhance their performance with comparably low cost. Many of these methods have been developed and evaluated on mathematical reasoning benchmarks such as AIME. This paper investigates whether the lessons learned from these benchmarks generalize to the domain of advanced theoretical physics. We evaluate a range of common test-time scaling methods on the TPBench physics dataset and compare their effectiveness with results on AIME. To better leverage the structure of physics problems, we develop a novel, symbolic weak-verifier framework to improve parallel scaling results. Our empirical results demonstrate that this method significantly outperforms existing test-time scaling approaches on TPBench. We also evaluate our method on AIME, confirming its effectiveness in solving advanced mathematical problems. Our findings highlight the power of step-wise symbolic verification for tackling complex scientific problems.