chunhuizhang / llm_rlLinks
llm & rl
☆209Updated this week
Alternatives and similar repositories for llm_rl
Users that are interested in llm_rl are comparing it to the libraries listed below
Sorting:
- 在verl上做reward的定制开发☆111Updated 3 months ago
- Latest Advances on Long Chain-of-Thought Reasoning☆497Updated last month
- ☆386Updated last week
- Agent-R1: Training Powerful LLM Agents with End-to-End Reinforcement Learning☆789Updated last month
- ☆374Updated 7 months ago
- An Awesome List of Agentic Model trained with Reinforcement Learning☆461Updated last week
- Awesome Agent Training☆223Updated last week
- Awesome RL-based LLM Reasoning☆613Updated last month
- WWW2025 Multimodal Intent Recognition for Dialogue Systems Challenge☆126Updated 10 months ago
- ☆330Updated 3 months ago
- ☆546Updated 8 months ago
- LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment☆372Updated last year
- OpenRFT: Adapting Reasoning Foundation Model for Domain-specific Tasks with Reinforcement Fine-Tuning☆148Updated 8 months ago
- Official code for the paper, "Stop Summation: Min-Form Credit Assignment Is All Process Reward Model Needs for Reasoning"☆136Updated last month
- A live reading list for LLM data synthesis (Updated to July, 2025).☆370Updated 2 weeks ago
- Generative AI Act II: Test Time Scaling Drives Cognition Engineering☆205Updated 4 months ago
- A series of technical report on Slow Thinking with LLM☆727Updated last month
- a-m-team's exploration in large language modeling☆187Updated 3 months ago
- The related works and background techniques about Openai o1☆224Updated 8 months ago
- ☆283Updated 3 months ago
- 🔧Tool-Star: Empowering LLM-brained Multi-Tool Reasoner via Reinforcement Learning☆252Updated last week
- Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement fine-tuning (RFT) of large language models (…☆330Updated this week
- ☆106Updated 11 months ago
- minimal-cost for training 0.5B R1-Zero☆768Updated 3 months ago
- ☆47Updated 7 months ago
- 这是一个open-r1的复现项目,对0.5B、1.5B、3B、7B的qwen模型进行GRPO训练,观察到一些有趣的现象。☆44Updated 5 months ago
- ☆264Updated 2 months ago
- ☆102Updated last year
- Official Repository of "Learning to Reason under Off-Policy Guidance"☆295Updated this week
- ☆97Updated 3 months ago