huggingface / transformersLinks
π€ Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
β151,997Updated this week
Alternatives and similar repositories for transformers
Users that are interested in transformers are comparing it to the libraries listed below
Sorting:
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β31,918Updated last month
- Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.β30,349Updated this week
- π€ The largest hub of ready-to-use datasets for AI models with fast, easy-to-use and efficient data manipulation toolsβ20,801Updated this week
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesβ21,798Updated 4 months ago
- State-of-the-Art Text Embeddingsβ17,774Updated 2 weeks ago
- TensorFlow code and pre-trained models for BERTβ39,619Updated last year
- A library for efficient similarity search and clustering of dense vectors.β37,735Updated this week
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β19,959Updated this week
- Tensors and Dynamic neural networks in Python with strong GPU accelerationβ94,387Updated this week
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Autoβ¦β16,017Updated this week
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) trainingβ22,853Updated last year
- Build and share delightful machine learning apps, all in Python. π Star to support our work!β40,389Updated this week
- Ongoing research training transformer models at scaleβ13,976Updated last week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β40,538Updated last week
- State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterβ¦β14,542Updated last year
- The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights --β¦β35,625Updated this week
- Fast and memory-efficient exact attentionβ20,280Updated this week
- π₯ Fast State-of-the-Art Tokenizers optimized for Research and Productionβ10,196Updated 2 weeks ago
- Google Researchβ36,659Updated this week
- Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and moreβ33,852Updated this week
- π« Industrial-strength Natural Language Processing (NLP) in Pythonβ32,758Updated last week
- An annotated implementation of the Transformer paper.β6,678Updated last year
- A framework for training and evaluating AI models on a variety of openly available dialogue datasets.β10,623Updated 2 years ago
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β9,256Updated 2 weeks ago
- Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.β16,650Updated 2 years ago
- Interactive deep learning book with multi-framework code, math, and discussions. Adopted at 500 universities from 70 countries including β¦β27,223Updated last year
- Making large AI models cheaper, faster and more accessibleβ41,220Updated 3 weeks ago
- Visualizer for neural network, deep learning and machine learning modelsβ31,672Updated this week
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β12,885Updated 10 months ago
- An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.β39,199Updated 5 months ago