🤖 AI Summary
This work identifies two root causes of Proximal Policy Optimization (PPO) failure in long-chain-of-thought (Long-CoT) reasoning: (1) value function initialization bias and (2) severe reward signal attenuation along the reasoning chain. To address these, we propose Value-Calibrated PPO (VC-PPO), featuring: (1) a pre-trained specialized value model to mitigate initialization bias, and (2) a decoupled generalized advantage estimation (GAE) scheme that computes advantages independently of the value prediction trajectory, thereby suppressing error propagation. VC-PPO jointly optimizes reinforcement learning and chain-of-thought generation. On the AIME mathematical reasoning benchmark, VC-PPO significantly improves PPO’s success rate. Ablation studies confirm that both components are critical for training stability. This is the first work to systematically diagnose and resolve the fundamental failure mechanisms of PPO in Long-CoT settings.
📝 Abstract
Reinforcement learning (RL) is pivotal for enabling large language models (LLMs) to generate long chains of thought (CoT) for complex tasks like math and reasoning. However, Proximal Policy Optimization (PPO), effective in many RL scenarios, fails in long CoT tasks. This paper identifies that value initialization bias and reward signal decay are the root causes of PPO's failure. We propose Value-Calibrated PPO (VC-PPO) to address these issues. In VC-PPO, the value model is pretrained to tackle initialization bias, and the Generalized Advantage Estimation (GAE) computation is decoupled between the actor and critic to mitigate reward signal decay. Experiments on the American Invitational Mathematics Examination (AIME) show that VC-PPO significantly boosts PPO performance. Ablation studies show that techniques in VC-PPO are essential in enhancing PPO for long CoT tasks.