Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world.
β2,402Jan 11, 2026Updated 2 months ago
Alternatives and similar repositories for hivemind
Users that are interested in hivemind are comparing it to the libraries listed below
Sorting:
- Official code for "Distributed Deep Learning in Open Collaborations" (NeurIPS 2021)β118Jan 13, 2022Updated 4 years ago
- πΈ Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloadingβ10,003Sep 7, 2024Updated last year
- β399Jan 31, 2026Updated last month
- Official code for "SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient"β149Dec 11, 2023Updated 2 years ago
- "Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts" (NeurIPS 2020), original PyTorch implemenβ¦β56Nov 5, 2020Updated 5 years ago
- Training a model similar to OpenAI DALL-E with volunteers from all over the Internet using hivemind and dalle-pytorch (NeurIPS 2021 demo)β27May 29, 2023Updated 2 years ago
- Memory-efficient transformer. Work in progress.β19Sep 17, 2022Updated 3 years ago
- PyTorch extensions for high performance and large scale training.β3,403Apr 26, 2025Updated 10 months ago
- "Moshpit SGD: Communication-Efficient Decentralized Training on Heterogeneous Unreliable Devices", official implementationβ30Feb 4, 2025Updated last year
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed librariesβ7,399Feb 3, 2026Updated last month
- prime is a framework for efficient, globally distributed training of AI models over the internet.β852Nov 16, 2025Updated 4 months ago
- Accessible large language models via k-bit quantization for PyTorch.β8,052Updated this week
- Efficient Deep Learning Systems course materials (HSE, YSDA)β965Updated this week
- Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.β30,926Mar 10, 2026Updated last week
- OpenDiLoCo: An Open-Source Framework for Globally Distributed Low-Communication Trainingβ562Jan 13, 2025Updated last year
- A Smart, Automatic, Fast and Lightweight Web Scraper for Pythonβ7,123Jun 9, 2025Updated 9 months ago
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β9,563Updated this week
- Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)β9,430Feb 20, 2026Updated last month
- Running large language models on a single GPU for throughput-oriented scenarios.β9,380Oct 28, 2024Updated last year
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.β1,077Apr 17, 2024Updated last year
- Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 20+ clouds, oβ¦β9,576Updated this week
- Development repository for the Triton language and compilerβ18,656Updated this week
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)β¦β14,419Mar 5, 2026Updated 2 weeks ago
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β41,807Mar 13, 2026Updated last week
- a libp2p-backed daemon wrapping the functionalities of go-libp2p for use in other languagesβ11Feb 9, 2025Updated last year
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)β4,739Jan 8, 2024Updated 2 years ago
- Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and moreβ35,108Updated this week
- Accelerated deep learning R&Dβ3,371Jun 27, 2025Updated 8 months ago
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (Nβ¦β4,709Updated this week
- Library for 8-bit optimizers and quantization routines.β780Aug 18, 2022Updated 3 years ago
- Fast and memory-efficient exact attentionβ22,832Updated this week
- Go ahead and axolotl questionsβ11,460Updated this week
- functorch is JAX-like composable function transforms for PyTorch.β1,437Aug 21, 2025Updated 6 months ago
- A data augmentations library for audio, image, text, and video.β5,070Updated this week
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) trainingβ23,919Aug 15, 2024Updated last year
- Incentivized Training over Wide Web with 1000x model compression.β22Oct 30, 2024Updated last year
- Type annotations and dynamic checking for a tensor's shape, dtype, names, etc.β1,475May 2, 2025Updated 10 months ago
- Lightning Training strategy for HiveMindβ18Jan 20, 2026Updated 2 months ago
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β32,191Sep 30, 2025Updated 5 months ago