learning-at-home / hivemindView external linksLinks
Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world.
β2,374Jan 11, 2026Updated last month
Alternatives and similar repositories for hivemind
Users that are interested in hivemind are comparing it to the libraries listed below
Sorting:
- πΈ Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloadingβ9,930Sep 7, 2024Updated last year
- Official code for "Distributed Deep Learning in Open Collaborations" (NeurIPS 2021)β117Jan 13, 2022Updated 4 years ago
- Official code for "SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient"β149Dec 11, 2023Updated 2 years ago
- "Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts" (NeurIPS 2020), original PyTorch implemenβ¦β56Nov 5, 2020Updated 5 years ago
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed librariesβ7,383Feb 3, 2026Updated last week
- PyTorch extensions for high performance and large scale training.β3,397Apr 26, 2025Updated 9 months ago
- Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.β30,823Feb 4, 2026Updated last week
- Accessible large language models via k-bit quantization for PyTorch.β7,939Jan 22, 2026Updated 3 weeks ago
- Training a model similar to OpenAI DALL-E with volunteers from all over the Internet using hivemind and dalle-pytorch (NeurIPS 2021 demo)β27May 29, 2023Updated 2 years ago
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β9,491Feb 6, 2026Updated last week
- prime is a framework for efficient, globally distributed training of AI models over the internet.β850Nov 16, 2025Updated 2 months ago
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)β¦β14,351Updated this week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β41,578Feb 7, 2026Updated last week
- Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)β9,395Jan 26, 2026Updated 2 weeks ago
- Memory-efficient transformer. Work in progress.β19Sep 17, 2022Updated 3 years ago
- A data augmentations library for audio, image, text, and video.β5,067Updated this week
- Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and moreβ34,848Updated this week
- Running large language models on a single GPU for throughput-oriented scenarios.β9,384Oct 28, 2024Updated last year
- Development repository for the Triton language and compilerβ18,387Updated this week
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)β4,742Jan 8, 2024Updated 2 years ago
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (Nβ¦β4,704Jan 12, 2026Updated last month
- Your PyTorch AI Factory - Flash enables you to easily configure and run complex AI recipes for over 15 tasks across 7 data domainsβ1,736Oct 8, 2023Updated 2 years ago
- "Moshpit SGD: Communication-Efficient Decentralized Training on Heterogeneous Unreliable Devices", official implementationβ30Feb 4, 2025Updated last year
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Autoβ¦β16,738Updated this week
- Accelerated deep learning R&Dβ3,372Jun 27, 2025Updated 7 months ago
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β32,143Sep 30, 2025Updated 4 months ago
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) trainingβ23,494Aug 15, 2024Updated last year
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.β1,072Apr 17, 2024Updated last year
- Database for AI. Store Vectors, Images, Texts, Videos, etc. Use with LLMs/LangChain. Store, query, version, & visualize any AI data. Streβ¦β8,993Feb 7, 2026Updated last week
- An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.β8,287Feb 25, 2022Updated 3 years ago
- Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 20+ clouds, oβ¦β9,442Updated this week
- Fast and memory-efficient exact attentionβ22,231Updated this week
- OpenDiLoCo: An Open-Source Framework for Globally Distributed Low-Communication Trainingβ562Jan 13, 2025Updated last year
- Sparsity-aware deep learning inference runtime for CPUsβ3,161Jun 2, 2025Updated 8 months ago
- Model parallel transformers in JAX and Haikuβ6,364Jan 21, 2023Updated 3 years ago
- Go ahead and axolotl questionsβ11,289Updated this week
- Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller modelsβ2,144Jun 2, 2025Updated 8 months ago
- You like pytorch? You like micrograd? You love tinygrad! β€οΈβ31,345Updated this week
- Training and serving large-scale neural networks with auto parallelization.β3,180Dec 9, 2023Updated 2 years ago