PrimeIntellect-ai / OpenDiloco
OpenDiLoCo: An Open-Source Framework for Globally Distributed Low-Communication Training
☆498Updated 4 months ago
Alternatives and similar repositories for OpenDiloco
Users that are interested in OpenDiloco are comparing it to the libraries listed below
Sorting:
- prime is a framework for efficient, globally distributed training of AI models over the internet.☆743Updated last week
- Distributed Training Over-The-Internet☆920Updated 5 months ago
- prime-rl is a codebase for decentralized RL training at scale☆211Updated this week
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆327Updated 5 months ago
- VPTQ, A Flexible and Extreme low-bit quantization algorithm☆633Updated 3 weeks ago
- Muon is Scalable for LLM Training☆1,044Updated last month
- [ICML 2024] CLLMs: Consistency Large Language Models☆391Updated 6 months ago
- Efficient LLM Inference over Long Sequences☆373Updated 2 weeks ago
- OLMoE: Open Mixture-of-Experts Language Models☆746Updated 2 months ago
- noise_step: Training in 1.58b With No Gradient Memory☆219Updated 4 months ago
- Pretraining code for a large-scale depth-recurrent language model☆760Updated last month
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.☆723Updated 7 months ago
- A throughput-oriented high-performance serving framework for LLMs☆806Updated last week
- Minimalistic large language model 3D-parallelism training☆1,870Updated this week
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆233Updated 3 months ago
- procedural reasoning datasets☆580Updated this week
- Single File, Single GPU, From Scratch, Efficient, Full Parameter Tuning library for "RL for LLMs"☆450Updated this week
- Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch☆512Updated 6 months ago
- Long context evaluation for large language models☆208Updated 2 months ago
- DeMo: Decoupled Momentum Optimization☆186Updated 5 months ago
- [NeurIPS'24 Spotlight, ICLR'25, ICML'25] To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention…☆1,013Updated last week
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆199Updated 10 months ago
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems☆324Updated last week
- Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse …☆357Updated this week
- LLM KV cache compression made easy☆481Updated last week
- PyTorch per step fault tolerance (actively under development)☆300Updated this week
- PyTorch building blocks for the OLMo ecosystem☆212Updated this week
- GRadient-INformed MoE☆262Updated 7 months ago
- ☆186Updated 3 months ago
- ☆534Updated 8 months ago