sunblaze-ucb / IntuitorLinks
Code for the paper: "Learning to Reason without External Rewards"
☆295Updated last week
Alternatives and similar repositories for Intuitor
Users that are interested in Intuitor are comparing it to the libraries listed below
Sorting:
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆219Updated last month
- ☆114Updated 5 months ago
- official repository for “Reinforcement Learning for Reasoning in Large Language Models with One Training Example”☆290Updated this week
- ☆292Updated last week
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate"☆159Updated 2 weeks ago
- ☆207Updated 4 months ago
- ☆300Updated 3 weeks ago
- ☆190Updated 2 months ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆222Updated last month
- ☆203Updated 4 months ago
- RL Scaling and Test-Time Scaling (ICML'25)☆106Updated 5 months ago
- Repo of paper "Free Process Rewards without Process Labels"☆153Updated 3 months ago
- SkyRL-v0: Train Real-World Long-Horizon Agents via Reinforcement Learning☆422Updated this week
- Tina: Tiny Reasoning Models via LoRA☆260Updated 3 weeks ago
- ☆180Updated 2 months ago
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"☆238Updated last month
- ☆169Updated this week
- ☆119Updated last month
- ☆157Updated 3 weeks ago
- A Large-Scale, Challenging, Decontaminated, and Verifiable Mathematical Dataset for Advancing Reasoning☆220Updated 2 weeks ago
- Resources for our paper: "Agent-R: Training Language Model Agents to Reflect via Iterative Self-Training"☆149Updated 2 weeks ago
- A version of verl to support tool use☆251Updated this week
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆94Updated 3 months ago
- 🌾 OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.☆379Updated 2 weeks ago
- ☆220Updated last month
- Code and example data for the paper: Rule Based Rewards for Language Model Safety☆188Updated 11 months ago
- repo for paper https://arxiv.org/abs/2504.13837☆158Updated last month
- General Reasoner: Advancing LLM Reasoning Across All Domains☆141Updated 2 weeks ago
- ☆115Updated 4 months ago
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆93Updated 2 weeks ago