SiliangZeng / Multi-Turn-RL-AgentLinks
☆100Updated 6 months ago
Alternatives and similar repositories for Multi-Turn-RL-Agent
Users that are interested in Multi-Turn-RL-Agent are comparing it to the libraries listed below
Sorting:
- official implementation of paper "Process Reward Model with Q-value Rankings"☆65Updated 10 months ago
- ☆116Updated 11 months ago
- RL Scaling and Test-Time Scaling (ICML'25)☆112Updated 11 months ago
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆254Updated 7 months ago
- Critique-out-Loud Reward Models☆70Updated last year
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples☆112Updated 5 months ago
- ☆50Updated 10 months ago
- Repo of paper "Free Process Rewards without Process Labels"☆168Updated 9 months ago
- ☆213Updated 6 months ago
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆114Updated 4 months ago
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"☆270Updated 2 months ago
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆86Updated 9 months ago
- Process Reward Models That Think☆67Updated last month
- A Large-Scale, High-Quality Math Dataset for Reinforcement Learning in Language Models☆70Updated 10 months ago
- Code for the paper: "Learning to Reason without External Rewards"☆383Updated 5 months ago
- A repo for open research on building large reasoning models☆125Updated last week
- [ICML 2025] Teaching Language Models to Critique via Reinforcement Learning☆118Updated 7 months ago
- Interpretable Contrastive Monte Carlo Tree Search Reasoning☆48Updated last year
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆134Updated 9 months ago
- Natural Language Reinforcement Learning☆101Updated 4 months ago
- ☆345Updated 5 months ago
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆63Updated last year
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate" [COLM 2025]☆180Updated 5 months ago
- General Reasoner: Advancing LLM Reasoning Across All Domains [NeurIPS25]☆211Updated last month
- ☆105Updated 2 weeks ago
- Revisiting Mid-training in the Era of Reinforcement Learning Scaling☆182Updated 5 months ago
- ☆106Updated 2 weeks ago
- ☆86Updated last month
- ☆68Updated 6 months ago
- B-STAR: Monitoring and Balancing Exploration and Exploitation in Self-Taught Reasoners☆86Updated 7 months ago