PrimeIntellect-ai / OpenDiloco
OpenDiLoCo: An Open-Source Framework for Globally Distributed Low-Communication Training
☆424Updated 2 weeks ago
Alternatives and similar repositories for OpenDiloco:
Users that are interested in OpenDiloco are comparing it to the libraries listed below
- prime is a framework for efficient, globally distributed training of AI models over the internet.☆626Updated this week
- Distributed Training Over-The-Internet☆866Updated last month
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆288Updated last month
- GRadient-INformed MoE☆261Updated 4 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆190Updated 6 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated 3 months ago
- Efficient LLM Inference over Long Sequences☆349Updated last month
- Long context evaluation for large language models☆198Updated this week
- VPTQ, A Flexible and Extreme low-bit quantization algorithm☆572Updated last week
- An Open Source Toolkit For LLM Distillation☆442Updated 3 weeks ago
- [ICML 2024] CLLMs: Consistency Large Language Models☆368Updated 2 months ago
- OLMoE: Open Mixture-of-Experts Language Models☆536Updated last month
- ☆192Updated this week
- Training Large Language Model to Reason in a Continuous Latent Space☆746Updated this week
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆263Updated last year
- Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM☆894Updated this week
- DeMo: Decoupled Momentum Optimization☆171Updated last month
- ☆497Updated 5 months ago
- ☆517Updated 3 months ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆249Updated 6 months ago
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆394Updated 4 months ago
- noise_step: Training in 1.58b With No Gradient Memory☆215Updated last month
- A high-throughput and memory-efficient inference and serving engine for LLMs☆257Updated 3 months ago
- Minimalistic 4D-parallelism distributed training framework for education purpose☆670Updated this week
- A Self-adaptation Framework🐙 that adapts LLMs for unseen tasks in real-time!☆831Updated 2 weeks ago
- Official implementation of Half-Quadratic Quantization (HQQ)☆737Updated 2 weeks ago
- Efficient, Flexible and Portable Structured Generation☆619Updated this week
- A throughput-oriented high-performance serving framework for LLMs☆714Updated 4 months ago
- Synthetic Data curation for post-training and structured data extraction☆575Updated this week