π€ AI Summary
This paper addresses the fundamental mismatch between the training objective (Chain-of-Thought, CoT, risk) and the test objective (end-to-end risk) under CoT supervision. We establish the first statistical learning theory framework for CoT reasoning. Central to our approach is the novel information-theoretic measure βCoT information contentβ β_CoT(Ξ΅; β), which explicitly quantifies the relationship between CoT risk and end-to-end risk. Leveraging this measure, we derive a tight sample complexity upper bound of d / β_CoT(Ξ΅; β), substantially improving upon the standard supervised learning bound d / Ξ΅. We further prove an information-theoretic lower bound, rigorously establishing the optimality of our upper bound. Our analysis integrates information-theoretic techniques, risk decomposition, and hypothesis class complexity modeling. This work provides the first theoretical guarantee for the improved generalization of CoT-based reasoning and identifies CoT information content as the intrinsic statistical complexity measure governing this paradigm.
π Abstract
Learning complex functions that involve multi-step reasoning poses a significant challenge for standard supervised learning from input-output examples. Chain-of-thought (CoT) supervision, which provides intermediate reasoning steps together with the final output, has emerged as a powerful empirical technique, underpinning much of the recent progress in the reasoning capabilities of large language models. This paper develops a statistical theory of learning under CoT supervision. A key characteristic of the CoT setting, in contrast to standard supervision, is the mismatch between the training objective (CoT risk) and the test objective (end-to-end risk). A central part of our analysis, distinguished from prior work, is explicitly linking those two types of risk to achieve sharper sample complexity bounds. This is achieved via the *CoT information measure* $mathcal{I}_{mathcal{D}, h_star}^{mathrm{CoT}}(epsilon; calH)$, which quantifies the additional discriminative power gained from observing the reasoning process. The main theoretical results demonstrate how CoT supervision can yield significantly faster learning rates compared to standard E2E supervision. Specifically, it is shown that the sample complexity required to achieve a target E2E error $epsilon$ scales as $d/mathcal{I}_{mathcal{D}, h_star}^{mathrm{CoT}}(epsilon; calH)$, where $d$ is a measure of hypothesis class complexity, which can be much faster than standard $d/epsilon$ rates. Information-theoretic lower bounds in terms of the CoT information are also obtained. Together, these results suggest that CoT information is a fundamental measure of statistical complexity for learning under chain-of-thought supervision.