AIFrameResearch / SPOLinks
Segment Policy Optimization: Improved Credit Assignment in Reinforcement Learning for LLMs
☆27Updated 2 weeks ago
Alternatives and similar repositories for SPO
Users that are interested in SPO are comparing it to the libraries listed below
Sorting:
- Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".☆81Updated 2 months ago
- Official Repository of "Learning to Reason under Off-Policy Guidance"☆271Updated 3 weeks ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆237Updated 2 months ago
- ☆263Updated 2 months ago
- Extrapolating RLVR to General Domains without Verifiers☆136Updated last week
- Implementation for the research paper "Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision".☆56Updated 8 months ago
- The Entropy Mechanism of Reinforcement Learning for Large Language Model Reasoning.☆282Updated last month
- Tool-Star: Empowering LLM-brained Multi-Tool Reasoner via Reinforcement Learning☆225Updated last week
- xVerify: Efficient Answer Verifier for Reasoning Model Evaluations☆127Updated 3 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆82Updated 5 months ago
- Official code for the paper, "Stop Summation: Min-Form Credit Assignment Is All Process Reward Model Needs for Reasoning"☆133Updated 3 weeks ago
- Official repository for paper: O1-Pruner: Length-Harmonizing Fine-Tuning for O1-Like Reasoning Pruning☆86Updated 5 months ago
- Chain of Thoughts (CoT) is so hot! so long! We need short reasoning process!☆68Updated 4 months ago
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆100Updated this week
- ☆67Updated last month
- Official Repository of "Learning what reinforcement learning can't"☆54Updated this week
- [ACL' 25] The official code repository for PRMBench: A Fine-grained and Challenging Benchmark for Process-Level Reward Models.☆78Updated 5 months ago
- Source code for our paper: "ARIA: Training Language Agents with Intention-Driven Reward Aggregation".☆20Updated this week
- ☆323Updated last week
- This is the official implementation of the paper "S²R: Teaching LLMs to Self-verify and Self-correct via Reinforcement Learning"☆69Updated 3 months ago
- ☆197Updated last week
- ☆141Updated 2 months ago
- CPPO: Accelerating the Training of Group Relative Policy Optimization-Based Reasoning Models☆147Updated 2 months ago
- Test-time preferenece optimization (ICML 2025).☆158Updated 3 months ago
- Pre-trained, Scalable, High-performance Reward Models via Policy Discriminative Learning.☆146Updated last month
- A comprehensive collection of process reward models.☆99Updated 2 weeks ago
- repo for paper https://arxiv.org/abs/2504.13837☆180Updated last month
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆76Updated 4 months ago
- ☆46Updated 4 months ago
- A unified suite for generating elite reasoning problems and training high-performance LLMs, including pioneering attention-free architect…☆64Updated 2 months ago