A Survey of Reinforcement Learning for Large Reasoning Models
A Survey of Reinforcement Learning for Large Reasoning Models
Paper: arXiv:2509.08827
Authors: Kaiyan Zhang, Yuxin Zuo, Bingxiang He, et al. (2025)
1. Introduction
Large Reasoning Models (LRMs) extend LLMs by focusing on complex problem-solving and multi-step reasoning.
This survey explores how Reinforcement Learning (RL) is reshaping LRMs, what components are essential, what challenges remain, and where the field is heading.
2. Foundational Components of RL for LLMs
Reward Design
- 
Verifiable rewards
These rewards are based on tasks where correctness can be objectively checked.- Examples include verifying a math solution, running unit tests for generated code, or checking if output formatting matches specifications.
 - Strength: reliable and automated.
 - Limitation: not every reasoning task is easily verifiable.
 
 - 
Learned reward models
When ground-truth verification is not possible (e.g., evaluating free-form text quality), a separate reward model is trained.- This model scores outputs based on human feedback or preferences.
 - Useful for subjective or open-ended tasks.
 - Risk: reward models can encode biases or be gamed by the LLM (“reward hacking”).
 
 - 
Dense & shaped rewards
Instead of only rewarding the final outcome, shaped rewards give intermediate feedback at each reasoning step.- Example: partial credit for writing the correct equation even if the final math answer is wrong.
 - Benefit: faster and more stable learning.
 - Challenge: designing meaningful step-level rewards is task-dependent and labor-intensive.
 
 - 
Unsupervised signals
Rewards can also come from intrinsic measures, without explicit labels.- Examples: promoting output diversity, penalizing repetition, or rewarding lower entropy.
 - Advantage: requires no manual labeling.
 - Drawback: weak alignment with actual task objectives.
 
 
Policy Optimization
- 
Policy gradient methods (REINFORCE, PPO)
These directly adjust the model’s parameters to maximize expected reward.- REINFORCE is simple but noisy.
 - PPO (Proximal Policy Optimization) is widely used for LLMs because it balances improvement with stability by limiting step size.
 
 - 
Critic-based approaches (actor-critic)
In this setup, the actor generates outputs while a critic estimates their value.- Advantage: more sample-efficient and stable.
 - Widely used in long reasoning tasks where variance is high.
 
 - 
Critic-free optimization
Eliminates the critic component, relying only on policy gradients and heuristic baselines.- Simpler and cheaper, but more prone to instability.
 
 - 
Off-policy methods
These allow training on previously collected trajectories rather than requiring live interaction.- Example: reusing logs of model generations.
 - Key for scaling, since collecting fresh samples from large models is very expensive.
 
 - 
Regularization for stability and diversity
To avoid collapse (e.g., repeating the same safe outputs), additional penalties are introduced.- KL divergence regularization ensures outputs don’t drift too far from the base model.
 - Entropy bonuses encourage exploration and diversity in reasoning.
 
 
Sampling Strategies
- 
Structured vs. dynamic sampling
- Structured: pre-defined, fixed rollout numbers and task distributions.
 - Dynamic: adaptive sampling that focuses more on difficult tasks or uncertain cases.
 
 - 
Balancing exploration vs. exploitation
- Exploration encourages the model to try novel reasoning paths.
 - Exploitation reinforces what is already working well.
 - Proper balance ensures both creativity and reliability.
 
 - 
Hyperparameter tuning for long reasoning chains
Parameters liketemperature,top-k, andtop-pstrongly affect output diversity.- Long reasoning chains require careful control of sampling to avoid drifting into irrelevant paths.
 - Rollout length also matters: too short may cut off reasoning, too long may add noise.
 
 
3. Foundational Problems
- 
RL vs. Supervised Fine-Tuning (SFT):
RL sharpens reasoning beyond SFT, but SFT often provides stronger priors. - 
Process vs. Outcome Rewards:
Process rewards encourage correct steps; outcome rewards only care about the final answer. - 
Model Priors:
RL interacts with pretraining biases and can either strengthen or override them. - 
Training Recipes:
Subtle choices (batch size, rollout strategy, reward shaping) drastically affect stability. 
4. Training Resources
- Static corpora: benchmark-style datasets, labeled corpora
 - Dynamic environments: interactive tasks, agents, simulators
 - Infrastructure: large-scale GPU/TPU clusters, verifier systems, scalable logging
 
5. Applications
RL for reasoning has been tested across multiple domains:
- Coding (unit-test based verifiable rewards)
 - Agentic tasks (planning, decision-making)
 - Multimodal reasoning (vision + language)
 - Multi-agent coordination
 - Medical reasoning (safe, verifiable responses)
 - Robotics (real-world sequential decision-making)
 
6. Future Directions
The survey highlights promising avenues:
- Continual RL – enabling lifelong learning
 - Memory-based RL – better context and past trajectory use
 - Model-based RL – planning with environment models
 - Latent space reasoning – efficient inference in compressed representations
 - Scientific discovery – leveraging RL for hypothesis testing and reasoning
 - LLM + Architecture Co-Design – aligning model structures with RL algorithms
 
7. Key Takeaways
- RL enhances reasoning but introduces reward design and stability challenges.
 - Verifiable rewards are critical for math, coding, and factual tasks.
 - RL is most effective when combined with supervised pretraining + careful reward shaping.
 - Applications extend beyond NLP into robotics, healthcare, and multi-agent systems.
 - Future research will need to address scalability, cost, and robust evaluation.
 
References
- Paper PDF
 - Zhang et al., A Survey of Reinforcement Learning for Large Reasoning Models, 2025