π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
β9,563Mar 17, 2026Updated this week
Alternatives and similar repositories for accelerate
Users that are interested in accelerate are comparing it to the libraries listed below
Sorting:
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β20,809Updated this week
- π Accelerate inference and training of π€ Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimizationβ¦β3,332Mar 13, 2026Updated last week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β41,869Updated this week
- π€ Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch.β33,085Updated this week
- Fast and memory-efficient exact attentionβ22,832Updated this week
- Hackable and optimized Transformers building blocks, supporting a composable construction.β10,373Updated this week
- Train transformer language models with reinforcement learning.β17,697Updated this week
- Accessible large language models via k-bit quantization for PyTorch.β8,052Updated this week
- Ongoing research training transformer models at scaleβ15,744Updated this week
- Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.β30,926Mar 10, 2026Updated last week
- PyTorch extensions for high performance and large scale training.β3,403Apr 26, 2025Updated 10 months ago
- π€ The largest hub of ready-to-use datasets for AI models with fast, easy-to-use and efficient data manipulation toolsβ21,289Updated this week
- π€ Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal modelβ¦β158,060Updated this week
- Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)β9,430Feb 20, 2026Updated last month
- Development repository for the Triton language and compilerβ18,708Updated this week
- π€ Evaluate: A library for easily evaluating machine learning models and datasets.β2,429Mar 10, 2026Updated last week
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β32,190Sep 30, 2025Updated 5 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMsβ73,479Updated this week
- A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorchβ8,936Updated this week
- The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights --β¦β36,504Mar 13, 2026Updated last week
- Large Language Model Text Generation Inferenceβ10,812Jan 8, 2026Updated 2 months ago
- π₯ Fast State-of-the-Art Tokenizers optimized for Research and Productionβ10,529Feb 28, 2026Updated 3 weeks ago
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesβ22,046Jan 23, 2026Updated last month
- Transformer related optimization, including BERT, GPTβ6,397Mar 27, 2024Updated last year
- Simple, safe way to store and distribute tensorsβ3,660Mar 12, 2026Updated last week
- CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an imageβ32,861Feb 18, 2026Updated last month
- An open source implementation of CLIP.β13,528Mar 12, 2026Updated last week
- Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and moreβ35,108Updated this week
- A library for efficient similarity search and clustering of dense vectors.β39,403Updated this week
- State-of-the-Art Text Embeddingsβ18,427Mar 12, 2026Updated last week
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Autoβ¦β16,918Updated this week
- SGLang is a high-performance serving framework for large language models and multimodal models.β24,829Updated this week
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β13,351Dec 17, 2024Updated last year
- Making large AI models cheaper, faster and more accessibleβ41,362Updated this week
- Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.β41,799Updated this week
- A PyTorch native platform for training generative AI modelsβ5,162Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMsβ10,858Jun 10, 2024Updated last year
- Unsupervised text tokenizer for Neural Network-based text generation.β11,700Mar 1, 2026Updated 2 weeks ago
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)β4,739Jan 8, 2024Updated 2 years ago