sail-sg / understand-r1-zeroLinks
Understanding R1-Zero-Like Training: A Critical Perspective
☆1,203Updated 5 months ago
Alternatives and similar repositories for understand-r1-zero
Users that are interested in understand-r1-zero are comparing it to the libraries listed below
Sorting:
- [COLM 2025] LIMO: Less is More for Reasoning☆1,061Updated 6 months ago
- 🌾 OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.☆623Updated last week
- ☆1,084Updated 3 weeks ago
- Official Repo for Open-Reasoner-Zero☆2,086Updated 8 months ago
- Recipes to scale inference-time compute of open models☆1,124Updated 8 months ago
- Large Reasoning Models☆807Updated last year
- An Open-source RL System from ByteDance Seed and Tsinghua AIR☆1,715Updated 8 months ago
- SkyRL: A Modular Full-stack RL Library for LLMs☆1,518Updated this week
- Scalable RL solution for advanced reasoning of language models☆1,803Updated 10 months ago
- Single File, Single GPU, From Scratch, Efficient, Full Parameter Tuning library for "RL for LLMs"☆589Updated 3 months ago
- [NeurIPS 2025] TTRL: Test-Time Reinforcement Learning☆972Updated 4 months ago
- ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search (NeurIPS 2024)☆688Updated last year
- A bibliography and survey of the papers surrounding o1☆1,213Updated last year
- Unleashing the Power of Reinforcement Learning for Math and Code Reasoners☆741Updated 7 months ago
- A series of technical report on Slow Thinking with LLM☆758Updated 5 months ago
- A version of verl to support diverse tool use☆852Updated 3 weeks ago
- ☆971Updated last year
- O1 Replication Journey☆2,001Updated last year
- Training Large Language Model to Reason in a Continuous Latent Space☆1,491Updated 5 months ago
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆944Updated 11 months ago
- ☆1,385Updated 4 months ago
- OLMoE: Open Mixture-of-Experts Language Models☆965Updated 4 months ago
- [NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewards☆1,326Updated 2 weeks ago
- Muon is Scalable for LLM Training☆1,421Updated 6 months ago
- [NeurIPS 2025 Spotlight] ReasonFlux (long-CoT), ReasonFlux-PRM (process reward model) and ReasonFlux-Coder (code generation)☆516Updated 4 months ago
- ☆328Updated 8 months ago
- ☆814Updated 7 months ago
- Parallel Scaling Law for Language Model — Beyond Parameter and Inference Time Scaling☆468Updated 8 months ago
- ☆761Updated last month
- Minimal hackable GRPO implementation☆321Updated last year