PRIME-RL / Entropy-Mechanism-of-RLLinks
The Entropy Mechanism of Reinforcement Learning for Large Language Model Reasoning.
☆296Updated last month
Alternatives and similar repositories for Entropy-Mechanism-of-RL
Users that are interested in Entropy-Mechanism-of-RL are comparing it to the libraries listed below
Sorting:
- Official Repository of "Learning to Reason under Off-Policy Guidance"☆282Updated last month
- Official code for the paper, "Stop Summation: Min-Form Credit Assignment Is All Process Reward Model Needs for Reasoning"☆133Updated last month
- ☆271Updated 3 months ago
- ☆261Updated last month
- ☆206Updated 6 months ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆248Updated 3 months ago
- ☆325Updated 3 weeks ago
- A version of verl to support tool use☆333Updated last week
- A Framework for LLM-based Multi-Agent Reinforced Training and Inference☆208Updated this week
- ☆204Updated 4 months ago
- A comprehensive collection of process reward models.☆104Updated last month
- This repository contains a regularly updated paper list for LLMs-reasoning-in-latent-space.☆153Updated this week
- 😎 A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, Agent, and Beyond☆286Updated last week
- Implementation for the research paper "Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision".☆56Updated 8 months ago
- ☆159Updated 3 months ago
- Repo of paper "Free Process Rewards without Process Labels"☆161Updated 5 months ago
- Chain of Thoughts (CoT) is so hot! so long! We need short reasoning process!☆69Updated 4 months ago
- Official Repository of "Learning what reinforcement learning can't"☆59Updated last week
- Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".☆81Updated 2 months ago
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆127Updated 5 months ago
- xVerify: Efficient Answer Verifier for Reasoning Model Evaluations☆128Updated 4 months ago
- ☆204Updated last week
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆101Updated 2 weeks ago
- [NeurIPS 2024 Oral] Aligner: Efficient Alignment by Learning to Correct☆184Updated 7 months ago
- Reference implementation for Token-level Direct Preference Optimization(TDPO)☆146Updated 6 months ago
- OpenRFT: Adapting Reasoning Foundation Model for Domain-specific Tasks with Reinforcement Fine-Tuning☆148Updated 8 months ago
- ☆66Updated 4 months ago
- ☆312Updated 2 months ago
- Implementation for "Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs"☆376Updated 7 months ago
- End-to-End Reinforcement Learning for Multi-Turn Tool-Integrated Reasoning☆172Updated 2 weeks ago