PrimeIntellect-ai / prime-dilocoLinks
prime is a framework for efficient, globally distributed training of AI models over the internet.
☆850Updated last month
Alternatives and similar repositories for prime-diloco
Users that are interested in prime-diloco are comparing it to the libraries listed below
Sorting:
- OpenDiLoCo: An Open-Source Framework for Globally Distributed Low-Communication Training☆557Updated last year
- Distributed Training Over-The-Internet☆973Updated 3 months ago
- Async RL Training at Scale☆985Updated this week
- Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse …☆825Updated this week
- ☆949Updated 2 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆370Updated last year
- Minimalistic large language model 3D-parallelism training☆2,411Updated last month
- OLMoE: Open Mixture-of-Experts Language Models☆950Updated 3 months ago
- Pretraining and inference code for a large-scale depth-recurrent language model☆859Updated 2 weeks ago
- 🎯An accuracy-first, highly efficient quantization toolkit for LLMs, designed to minimize quality degradation across Weight-Only Quantiza…☆806Updated this week
- Checkpoint-engine is a simple middleware to update model weights in LLM inference engines☆888Updated this week
- An Open Source Toolkit For LLM Distillation☆819Updated 3 weeks ago
- VPTQ, A Flexible and Extreme low-bit quantization algorithm☆671Updated 8 months ago
- Muon is Scalable for LLM Training☆1,397Updated 5 months ago
- Minimalistic 4D-parallelism distributed training framework for education purpose☆1,947Updated 4 months ago
- ☆584Updated last year
- A Self-adaptation Framework🐙 that adapts LLMs for unseen tasks in real-time!☆1,179Updated 11 months ago
- [NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewards☆1,304Updated last month
- Efficient LLM Inference over Long Sequences☆393Updated 6 months ago
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆939Updated last month
- Official implementation of Half-Quadratic Quantization (HQQ)☆905Updated 3 weeks ago
- Official inference library for pre-processing of Mistral models☆846Updated last week
- Scalable toolkit for efficient model reinforcement☆1,227Updated this week
- Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM☆2,553Updated this week
- A throughput-oriented high-performance serving framework for LLMs☆936Updated 2 months ago
- Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.☆346Updated last year
- Where GPUs get cooked 👩🍳🔥☆347Updated 3 months ago
- LLM KV cache compression made easy☆749Updated last month
- [NeurIPS'24 Spotlight, ICLR'25, ICML'25] To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention…☆1,171Updated 3 months ago
- Deep learning for dummies. All the practical details and useful utilities that go into working with real models.☆829Updated 5 months ago