efficientscaling / Z1Links
Repo for "Z1: Efficient Test-time Scaling with Code"
☆59Updated last month
Alternatives and similar repositories for Z1
Users that are interested in Z1 are comparing it to the libraries listed below
Sorting:
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆70Updated 2 months ago
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate"☆151Updated last month
- General Reasoner: Advancing LLM Reasoning Across All Domains☆126Updated this week
- Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling☆102Updated 4 months ago
- ☆46Updated 3 months ago
- Code for "Reasoning to Learn from Latent Thoughts"☆104Updated 2 months ago
- FastCuRL: Curriculum Reinforcement Learning with Stage-wise Context Scaling for Efficient Training R1-like Reasoning Models☆49Updated this week
- RM-R1: Unleashing the Reasoning Potential of Reward Models☆97Updated this week
- [NeurIPS 2024] Can LLMs Learn by Teaching for Better Reasoning? A Preliminary Study☆49Updated 6 months ago
- ☆127Updated 3 weeks ago
- [ICML 2025] Teaching Language Models to Critique via Reinforcement Learning☆98Updated 3 weeks ago
- ☆104Updated last month
- Codebase for Instruction Following without Instruction Tuning☆34Updated 8 months ago
- This is the official implementation of the paper "S²R: Teaching LLMs to Self-verify and Self-correct via Reinforcement Learning"☆64Updated last month
- Code for ICLR 2025 Paper "What is Wrong with Perplexity for Long-context Language Modeling?"☆78Updated 2 weeks ago
- A Sober Look at Language Model Reasoning☆52Updated this week
- Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".☆73Updated last month
- ☆89Updated last week
- A version of verl to support tool use☆172Updated this week
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆213Updated 3 weeks ago
- This is an official implementation of the Reward rAnked Fine-Tuning Algorithm (RAFT), also known as iterative best-of-n fine-tuning or re…☆31Updated 8 months ago
- Scaling Computer-Use Grounding via UI Decomposition and Synthesis☆49Updated this week
- ☆231Updated last week
- What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆64Updated 3 months ago
- [ACL 2025] A Generalizable and Purely Unsupervised Self-Training Framework☆60Updated this week
- A Large-Scale, High-Quality Math Dataset for Reinforcement Learning in Language Models☆55Updated 3 months ago
- ☆105Updated 2 months ago
- ☆45Updated 3 months ago
- ☆113Updated 4 months ago
- Dynamic Cheatsheet: Test-Time Learning with Adaptive Memory☆61Updated last week