allenai / OLMo-coreLinks
PyTorch building blocks for the OLMo ecosystem
β785Updated this week
Alternatives and similar repositories for OLMo-core
Users that are interested in OLMo-core are comparing it to the libraries listed below
Sorting:
- A project to improve skills of large language modelsβ813Updated this week
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β627Updated 2 weeks ago
- Scalable toolkit for efficient model reinforcementβ1,307Updated this week
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β371Updated last year
- Reproducible, flexible LLM evaluationsβ338Updated 2 weeks ago
- OLMoE: Open Mixture-of-Experts Language Modelsβ967Updated 4 months ago
- Automatic evals for LLMsβ579Updated last month
- Async RL Training at Scaleβ1,044Updated this week
- Open-source framework for the research and development of foundation models.β752Updated this week
- β961Updated 3 months ago
- [NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewardsβ1,332Updated 3 weeks ago
- SkyRL: A Modular Full-stack RL Library for LLMsβ1,547Updated this week
- An extension of the nanoGPT repository for training small MOE models.β236Updated 11 months ago
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ469Updated 8 months ago
- PyTorch-native post-training at scaleβ613Updated this week
- Single File, Single GPU, From Scratch, Efficient, Full Parameter Tuning library for "RL for LLMs"β593Updated 4 months ago
- β1,088Updated last month
- β232Updated 2 months ago
- [ICLR 2026] Tina: Tiny Reasoning Models via LoRAβ319Updated 4 months ago
- Miles is an enterprise-facing reinforcement learning framework for LLM and VLM post-training, forked from and co-evolving with slime.β858Updated this week
- Scalable toolkit for efficient model alignmentβ852Updated 4 months ago
- Open-source release accompanying Gao et al. 2025β501Updated 2 months ago
- Recipes to scale inference-time compute of open modelsβ1,124Updated 8 months ago
- Pretraining and inference code for a large-scale depth-recurrent language modelβ863Updated last month
- Muon is Scalable for LLM Trainingβ1,426Updated 6 months ago
- β220Updated 3 months ago
- A Gym for Agentic LLMsβ444Updated 3 weeks ago
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.β334Updated 3 months ago
- A scalable asynchronous reinforcement learning implementation with in-flight weight updates.β361Updated last week
- A bibliography and survey of the papers surrounding o1β1,212Updated last year