NousResearch / atroposLinks
Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse environments
☆853Updated this week
Alternatives and similar repositories for atropos
Users that are interested in atropos are comparing it to the libraries listed below
Sorting:
- Async RL Training at Scale☆1,044Updated this week
- [NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewards☆1,332Updated 3 weeks ago
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆458Updated last year
- Lightly-reviewed collection of community environments☆212Updated this week
- rl from zero pretrain, can it be done? yes.☆286Updated 4 months ago
- ☆237Updated last month
- Super basic implementation (gist-like) of RLMs with REPL environments.☆636Updated last month
- System 2 Reasoning Link Collection☆870Updated 10 months ago
- An interface library for RL post training with environments.☆1,132Updated this week
- Testing baseline LLMs performance across various models☆336Updated this week
- Inference-time scaling for LLMs-as-a-judge.☆328Updated 3 months ago
- Code for Paper: Training Software Engineering Agents and Verifiers with SWE-Gym [ICML 2025]☆627Updated 6 months ago
- Aidan Bench attempts to measure <big_model_smell> in LLMs.☆318Updated 7 months ago
- Distributed Training Over-The-Internet☆975Updated 3 months ago
- prime is a framework for efficient, globally distributed training of AI models over the internet.☆850Updated 2 months ago
- ☆118Updated last week
- Exploring Applications of GRPO☆251Updated 5 months ago
- ☆137Updated 10 months ago
- Build your own visual reasoning model☆418Updated 3 weeks ago
- MLGym A New Framework and Benchmark for Advancing AI Research Agents☆583Updated 6 months ago
- ⚖️ Awesome LLM Judges ⚖️☆161Updated 9 months ago
- Pretraining and inference code for a large-scale depth-recurrent language model☆863Updated last month
- Open source interpretability artefacts for R1.☆170Updated 9 months ago
- Harbor is a framework for running agent evaluations and creating and using RL environments.☆600Updated this week
- ☆961Updated 3 months ago
- open source interpretability platform 🧠☆704Updated this week
- Automatic evals for LLMs☆579Updated last month
- smol models are fun too☆93Updated last year
- [NeurIPS 2025 D&B Spotlight] Scaling Data for SWE-agents☆551Updated this week
- SkyRL: A Modular Full-stack RL Library for LLMs☆1,547Updated this week