prime is a framework for efficient, globally distributed training of AI models over the internet.
☆851Nov 16, 2025Updated 4 months ago
Alternatives and similar repositories for prime-diloco
Users that are interested in prime-diloco are comparing it to the libraries listed below
Sorting:
- OpenDiLoCo: An Open-Source Framework for Globally Distributed Low-Communication Training☆560Jan 13, 2025Updated last year
- Solidity contracts for the decentralized Prime Network protocol☆26Jul 6, 2025Updated 8 months ago
- Async RL Training at Scale☆1,156Updated this week
- peer-to-peer compute and intelligence network that enables decentralized AI development at scale☆137Nov 10, 2025Updated 4 months ago
- Distributed Training Over-The-Internet☆984Oct 14, 2025Updated 5 months ago
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆148Sep 12, 2025Updated 6 months ago
- Asynchronous P2P communication backend for decentralized pipeline parallelism☆42Jun 9, 2025Updated 9 months ago
- Modded vLLM to run pipeline parallelism over public networks☆40May 20, 2025Updated 10 months ago
- Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world.☆2,402Jan 11, 2026Updated 2 months ago
- Minimalistic large language model 3D-parallelism training☆2,617Feb 19, 2026Updated last month
- ☆34Sep 10, 2024Updated last year
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆487Updated this week
- A PyTorch native platform for training generative AI models☆5,162Updated this week
- supporting pytorch FSDP for optimizers☆84Dec 8, 2024Updated last year
- Efficient Triton Kernels for LLM Training☆6,216Updated this week
- TOPLOC: is a novel method for verifiable inference that enables users to verify that LLM providers are using the correct model configurat…☆52Apr 14, 2025Updated 11 months ago
- Minimalistic 4D-parallelism distributed training framework for education purpose☆2,116Aug 26, 2025Updated 6 months ago
- AllenAI's post-training codebase☆3,629Updated this week
- Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.☆4,757Jul 18, 2025Updated 8 months ago
- This library supports evaluating disparities in generated image quality, diversity, and consistency between geographic regions.☆20Jun 3, 2024Updated last year
- An Open Source Toolkit For LLM Distillation☆891Mar 14, 2026Updated last week
- Pretraining Code for METAGENE-1☆70Jan 6, 2025Updated last year
- ☆137Mar 20, 2025Updated last year
- PyTorch native quantization and sparsity for training and inference☆2,739Updated this week
- DeMo: Decoupled Momentum Optimization☆198Dec 2, 2024Updated last year
- Tools for merging pretrained large language models.☆6,867Mar 15, 2026Updated last week
- This repository contains the experimental PyTorch native float8 training UX☆226Aug 1, 2024Updated last year
- Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse …☆914Updated this week
- Code for BLT research paper☆2,030Nov 3, 2025Updated 4 months ago
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆282Nov 24, 2025Updated 3 months ago
- SIMD quantization kernels☆93Sep 7, 2025Updated 6 months ago
- SGLang is a high-performance serving framework for large language models and multimodal models.☆24,829Updated this week
- Ring attention implementation with flash attention☆996Sep 10, 2025Updated 6 months ago
- FlashInfer: Kernel Library for LLM Serving☆5,145Mar 15, 2026Updated last week
- ☆48Jan 18, 2024Updated 2 years ago
- torch implementation of diloco☆22May 31, 2024Updated last year
- Mamba support for transformer lens☆19Sep 17, 2024Updated last year
- GRadient-INformed MoE☆264Sep 25, 2024Updated last year
- A bidirectional pipeline parallelism algorithm for computation-communication overlap in DeepSeek V3/R1 training.☆2,934Jan 14, 2026Updated 2 months ago