lzhxmu / CPPOLinks
CPPO: Accelerating the Training of Group Relative Policy Optimization-Based Reasoning Models
☆126Updated this week
Alternatives and similar repositories for CPPO
Users that are interested in CPPO are comparing it to the libraries listed below
Sorting:
- Rethinking RL Scaling for Vision Language Models: A Transparent, From-Scratch Framework and Comprehensive Evaluation Scheme☆129Updated last month
- ☆202Updated 3 months ago
- An Easy-to-use, Scalable and High-performance RLHF Framework designed for Multimodal Models.☆127Updated 2 months ago
- ✨✨R1-Reward: Training Multimodal Reward Model Through Stable Reinforcement Learning☆136Updated 3 weeks ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆215Updated 3 weeks ago
- repo for paper https://arxiv.org/abs/2504.13837☆144Updated 2 weeks ago
- The official code of "VL-Rethinker: Incentivizing Self-Reflection of Vision-Language Models with Reinforcement Learning"☆103Updated 2 weeks ago
- [Preprint 2025] Thinkless: LLM Learns When to Think☆133Updated this week
- SFT or RL? An Early Investigation into Training R1-Like Reasoning Large Vision-Language Models☆115Updated last month
- Official code for the paper, "Stop Summation: Min-Form Credit Assignment Is All Process Reward Model Needs for Reasoning"☆123Updated this week
- xVerify: Efficient Answer Verifier for Reasoning Model Evaluations☆106Updated last month
- The Entropy Mechanism of Reinforcement Learning for Large Language Model Reasoning.☆103Updated this week
- ☆77Updated 5 months ago
- A RLHF Infrastructure for Vision-Language Models☆176Updated 6 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆69Updated 3 months ago
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆103Updated last week
- A comprehensive collection of process reward models.☆88Updated 2 weeks ago
- MMR1: Advancing the Frontiers of Multimodal Reasoning☆159Updated 2 months ago
- ☆101Updated last month
- Official repository for paper: O1-Pruner: Length-Harmonizing Fine-Tuning for O1-Like Reasoning Pruning☆80Updated 3 months ago
- ☆131Updated 3 weeks ago
- ☆100Updated last week
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆147Updated 2 months ago
- Official Repository of "Learning to Reason under Off-Policy Guidance"☆212Updated this week
- ☆295Updated last week
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆95Updated 2 months ago
- Chain of Thoughts (CoT) is so hot! so long! We need short reasoning process!☆54Updated 2 months ago
- ☆231Updated last week
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆180Updated 2 months ago
- ☆169Updated this week