NovaSky-AI / SkyRLLinks
SkyRL: A Modular Full-stack RL Library for LLMs
β1,518Updated this week
Alternatives and similar repositories for SkyRL
Users that are interested in SkyRL are comparing it to the libraries listed below
Sorting:
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β623Updated last week
- Scalable toolkit for efficient model reinforcementβ1,293Updated this week
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β673Updated 10 months ago
- β1,084Updated 3 weeks ago
- Understanding R1-Zero-Like Training: A Critical Perspectiveβ1,203Updated 5 months ago
- [NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewardsβ1,326Updated 2 weeks ago
- β1,385Updated 4 months ago
- Async RL Training at Scaleβ1,034Updated this week
- A project to improve skills of large language modelsβ804Updated this week
- A bibliography and survey of the papers surrounding o1β1,213Updated last year
- An Open-source RL System from ByteDance Seed and Tsinghua AIRβ1,715Updated 8 months ago
- slime is an LLM post-training framework for RL Scaling.β3,571Updated last week
- A version of verl to support diverse tool useβ852Updated 3 weeks ago
- Code for Paper: Training Software Engineering Agents and Verifiers with SWE-Gym [ICML 2025]β625Updated 6 months ago
- Recipes to scale inference-time compute of open modelsβ1,124Updated 8 months ago
- β957Updated 3 months ago
- Miles is an enterprise-facing reinforcement learning framework for LLM and VLM post-training, forked from and co-evolving with slime.β830Updated this week
- Automatic evals for LLMsβ578Updated last month
- [COLM 2025] LIMO: Less is More for Reasoningβ1,061Updated 6 months ago
- RAGEN leverages reinforcement learning to train LLM reasoning agents in interactive, stochastic environments.β2,503Updated last week
- A Gym for Agentic LLMsβ439Updated 2 weeks ago
- Meta Agents Research Environments is a comprehensive platform designed to evaluate AI agents in dynamic, realistic scenarios. Unlike statβ¦β418Updated last week
- [NeurIPS 2025] TTRL: Test-Time Reinforcement Learningβ972Updated 4 months ago
- Large Reasoning Modelsβ807Updated last year
- Checkpoint-engine is a simple middleware to update model weights in LLM inference enginesβ902Updated this week
- PyTorch building blocks for the OLMo ecosystemβ763Updated this week
- β971Updated last year
- Training Large Language Model to Reason in a Continuous Latent Spaceβ1,491Updated 5 months ago
- Scalable toolkit for efficient model alignmentβ848Updated 3 months ago
- ΟΒ²-Bench: Evaluating Conversational Agents in a Dual-Control Environmentβ690Updated last week