huggingface / transformersLinks
π€ Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
β152,334Updated this week
Alternatives and similar repositories for transformers
Users that are interested in transformers are comparing it to the libraries listed below
Sorting:
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β31,937Updated last month
- Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.β30,430Updated this week
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesβ21,831Updated 4 months ago
- π€ The largest hub of ready-to-use datasets for AI models with fast, easy-to-use and efficient data manipulation toolsβ20,847Updated this week
- Build and share delightful machine learning apps, all in Python. π Star to support our work!β40,461Updated last week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β40,670Updated this week
- CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an imageβ31,564Updated last year
- TensorFlow code and pre-trained models for BERTβ39,661Updated last year
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β20,050Updated this week
- State-of-the-Art Text Embeddingsβ17,849Updated 3 weeks ago
- State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterβ¦β14,576Updated last year
- Inference code for Llama modelsβ58,924Updated 9 months ago
- π€ Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch.β31,617Updated this week
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β12,915Updated 11 months ago
- The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights --β¦β35,773Updated last week
- Unsupervised text tokenizer for Neural Network-based text generation.β11,428Updated last week
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β9,289Updated this week
- Code for the paper "Language Models are Unsupervised Multitask Learners"β24,363Updated last year
- Fast and memory-efficient exact attentionβ20,541Updated this week
- A library for efficient similarity search and clustering of dense vectors.β37,969Updated this week
- Visualizer for neural network, deep learning and machine learning modelsβ31,793Updated this week
- Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and moreβ33,975Updated this week
- Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.β16,712Updated 2 years ago
- Code and documentation to train Stanford's Alpaca models, and generate the data.β30,211Updated last year
- A latent text-to-image diffusion modelβ71,828Updated last year
- Ongoing research training transformer models at scaleβ14,168Updated this week
- Google Researchβ36,696Updated this week
- The simplest, fastest repository for training/finetuning medium-sized GPTs.β49,543Updated this week
- An open source implementation of CLIP.β12,916Updated last week
- BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)β7,745Updated 5 months ago