Chengsong-Huang / R-ZeroLinks
codes for R-Zero: Self-Evolving Reasoning LLM from Zero Data (https://www.arxiv.org/pdf/2508.05004)
☆656Updated last month
Alternatives and similar repositories for R-Zero
Users that are interested in R-Zero are comparing it to the libraries listed below
Sorting:
- The offical repo for "Parallel-R1: Towards Parallel Thinking via Reinforcement Learning"☆229Updated 2 weeks ago
- Code and implementations for the paper "AgentGym-RL: Training LLM Agents for Long-Horizon Decision Making through Multi-Turn Reinforcemen…☆460Updated last month
- Training teachers with reinforcement learning able to make LLMs learn how to reason for test time scaling.☆347Updated 4 months ago
- A Scientific Multimodal Foundation Model☆587Updated last month
- OpenCUA: Open Foundations for Computer-Use Agents☆531Updated 3 weeks ago
- Research code artifacts for Code World Model (CWM) including inference tools, reproducibility, and documentation.☆699Updated last month
- An Open-Source Large-Scale Reinforcement Learning Project for Search Agents☆471Updated 3 weeks ago
- A MemAgent framework that can be extrapolated to 3.5M, along with a training framework for RL training of any agent workflow.☆761Updated 3 months ago
- ☆215Updated 2 weeks ago
- Chain-of-Agents: End-to-End Agent Foundation Models via Multi-Agent Distillation and Agentic RL.☆481Updated last month
- ☆828Updated last month
- ☆246Updated last month
- ☆651Updated this week
- ☆219Updated 8 months ago
- ☆1,320Updated last month
- Official Code of Memento: Fine-tuning LLM Agents without Fine-tuning LLMs☆1,808Updated 3 weeks ago
- Tina: Tiny Reasoning Models via LoRA☆302Updated last month
- Scaling RL on advanced reasoning models☆620Updated last week
- Official Repository for "Glyph: Scaling Context Windows via Visual-Text Compression"☆356Updated this week
- [NeurIPS 2025] Thinkless: LLM Learns When to Think☆237Updated last month
- [NeurIPS 2025] Reinforcement Learning for Reasoning in Large Language Models with One Training Example☆371Updated 2 weeks ago
- MCP-Universe is a comprehensive framework designed for developing, testing, and benchmarking AI agents☆474Updated last week
- Code for the paper: "Learning to Reason without External Rewards"☆369Updated 3 months ago
- Self-Adapting Language Models☆1,443Updated 3 months ago
- Next paradigm for LLM Agent. Unify plan and action through recursive code generation for adaptive, human-like decision-making.☆114Updated this week
- Official implementation of the NeurIPS 2025 paper "Soft Thinking: Unlocking the Reasoning Potential of LLMs in Continuous Concept Space"☆265Updated last month
- RLP: Reinforcement as a Pretraining Objective☆195Updated 3 weeks ago
- MCP-Bench: Benchmarking Tool-Using LLM Agents with Complex Real-World Tasks via MCP Servers☆364Updated 3 weeks ago
- Hypernetworks that adapt LLMs for specific benchmark tasks using only textual task description as the input☆905Updated 4 months ago
- [NeurIPS 2025] Atom of Thoughts for Markov LLM Test-Time Scaling☆591Updated 4 months ago