π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
β9,513Updated this week
Alternatives and similar repositories for accelerate
Users that are interested in accelerate are comparing it to the libraries listed below
Sorting:
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β20,678Updated this week
- π Accelerate inference and training of π€ Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimizationβ¦β3,296Feb 9, 2026Updated 2 weeks ago
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β41,648Updated this week
- Fast and memory-efficient exact attentionβ22,361Updated this week
- π€ Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch.β32,873Updated this week
- Hackable and optimized Transformers building blocks, supporting a composable construction.β10,353Feb 20, 2026Updated last week
- Accessible large language models via k-bit quantization for PyTorch.β7,997Updated this week
- Train transformer language models with reinforcement learning.β17,460Updated this week
- Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.β30,860Feb 21, 2026Updated last week
- Ongoing research training transformer models at scaleβ15,242Feb 21, 2026Updated last week
- PyTorch extensions for high performance and large scale training.β3,400Apr 26, 2025Updated 10 months ago
- π€ The largest hub of ready-to-use datasets for AI models with fast, easy-to-use and efficient data manipulation toolsβ21,228Updated this week
- Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)β9,401Feb 20, 2026Updated last week
- Development repository for the Triton language and compilerβ18,460Feb 22, 2026Updated last week
- π€ Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal modelβ¦β157,071Updated this week
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β32,170Sep 30, 2025Updated 5 months ago
- A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorchβ8,926Updated this week
- The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights --β¦β36,397Updated this week
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesβ22,033Jan 23, 2026Updated last month
- A high-throughput and memory-efficient inference and serving engine for LLMsβ71,234Updated this week
- Large Language Model Text Generation Inferenceβ10,774Jan 8, 2026Updated last month
- Transformer related optimization, including BERT, GPTβ6,394Mar 27, 2024Updated last year
- π€ Evaluate: A library for easily evaluating machine learning models and datasets.β2,419Jan 20, 2026Updated last month
- CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an imageβ32,642Feb 18, 2026Updated last week
- π₯ Fast State-of-the-Art Tokenizers optimized for Research and Productionβ10,485Updated this week
- An open source implementation of CLIP.β13,430Updated this week
- Simple, safe way to store and distribute tensorsβ3,637Updated this week
- Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and moreβ34,940Updated this week
- State-of-the-Art Text Embeddingsβ18,298Feb 20, 2026Updated last week
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Autoβ¦β16,807Updated this week
- A library for efficient similarity search and clustering of dense vectors.β39,195Updated this week
- SGLang is a high-performance serving framework for large language models and multimodal models.β23,658Updated this week
- Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.β41,516Updated this week
- Making large AI models cheaper, faster and more accessibleβ41,359Updated this week
- Flax is a neural network library for JAX that is designed for flexibility.β7,082Feb 22, 2026Updated last week
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β13,285Dec 17, 2024Updated last year
- A PyTorch native platform for training generative AI modelsβ5,098Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMsβ10,838Jun 10, 2024Updated last year
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.β24,478Aug 12, 2024Updated last year