Async RL Training at Scale
☆1,096Updated this week
Alternatives and similar repositories for prime-rl
Users that are interested in prime-rl are comparing it to the libraries listed below
Sorting:
- Our library for RL environments + evals☆3,869Updated this week
- Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse …☆872Updated this week
- A scalable asynchronous reinforcement learning implementation with in-flight weight updates.☆368Feb 19, 2026Updated last week
- [NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewards☆1,352Jan 16, 2026Updated last month
- Official CLI and Python SDK for Prime Intellect - access GPU compute, remote sandboxes, RL environments, and distributed training infrast…☆161Updated this week
- SkyRL: A Modular Full-stack RL Library for LLMs☆1,628Updated this week
- Scalable toolkit for efficient model reinforcement☆1,353Updated this week
- slime is an LLM post-training framework for RL Scaling.☆4,381Updated this week
- AllenAI's post-training codebase☆3,592Updated this week
- peer-to-peer compute and intelligence network that enables decentralized AI development at scale☆137Nov 10, 2025Updated 3 months ago
- Solidity contracts for the decentralized Prime Network protocol☆26Jul 6, 2025Updated 7 months ago
- Democratizing Reinforcement Learning for LLMs☆5,167Updated this week
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆143Sep 12, 2025Updated 5 months ago
- prime is a framework for efficient, globally distributed training of AI models over the internet.☆851Nov 16, 2025Updated 3 months ago
- 🌾 OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.☆633Jan 29, 2026Updated last month
- Scalable RL solution for advanced reasoning of language models☆1,809Mar 18, 2025Updated 11 months ago
- Lightly-reviewed collection of community environments☆215Updated this week
- A Gym for Agentic LLMs☆452Jan 21, 2026Updated last month
- Minimalistic large language model 3D-parallelism training☆2,579Feb 19, 2026Updated last week
- Build your own visual reasoning model☆419Jan 13, 2026Updated last month
- RAGEN leverages reinforcement learning to train LLM reasoning agents in interactive, stochastic environments.☆2,522Updated this week
- verl: Volcano Engine Reinforcement Learning for LLMs☆19,339Updated this week
- Train your own SOTA deductive reasoning model☆107Mar 6, 2025Updated 11 months ago
- Minimalistic 4D-parallelism distributed training framework for education purpose☆2,090Aug 26, 2025Updated 6 months ago
- General Reasoner: Advancing LLM Reasoning Across All Domains [NeurIPS25]☆221Nov 27, 2025Updated 3 months ago
- ☆1,104Jan 10, 2026Updated last month
- Automatic evals for LLMs☆580Updated this week
- The Automated LLM Speedrunning Benchmark measures how well LLM agents can reproduce previous innovations and discover new ones in languag…☆131Feb 21, 2026Updated last week
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆59Oct 18, 2025Updated 4 months ago
- Open-source framework for the research and development of foundation models.☆773Updated this week
- ☆19Mar 16, 2025Updated 11 months ago
- Checkpoint-engine is a simple middleware to update model weights in LLM inference engines☆912Updated this week
- Official Repo for Open-Reasoner-Zero☆2,087Jun 2, 2025Updated 8 months ago
- A PyTorch native platform for training generative AI models☆5,098Updated this week
- Parallel Scaling Law for Language Model — Beyond Parameter and Inference Time Scaling☆472May 17, 2025Updated 9 months ago
- Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.☆4,752Jul 18, 2025Updated 7 months ago
- Tools for merging pretrained large language models.☆6,814Jan 26, 2026Updated last month
- A Qwen .5B reasoning model trained on OpenR1-Math-220k☆14Oct 11, 2025Updated 4 months ago
- Single File, Single GPU, From Scratch, Efficient, Full Parameter Tuning library for "RL for LLMs"☆597Oct 7, 2025Updated 4 months ago