facebookresearch / fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
☆31,211Updated 2 months ago
Alternatives and similar repositories for fairseq:
Users that are interested in fairseq are comparing it to the libraries listed below
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities☆21,006Updated last month
- Unsupervised text tokenizer for Neural Network-based text generation.☆10,756Updated this week
- Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"☆6,311Updated last month
- A library for efficient similarity search and clustering of dense vectors.☆34,053Updated this week
- An open-source NLP research library, built on PyTorch.☆11,834Updated 2 years ago
- BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)☆7,283Updated last year
- 🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (i…☆8,568Updated this week
- 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.☆142,370Updated this week
- Ongoing research training transformer models at scale☆11,933Updated last week
- 🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.☆17,977Updated last week
- TensorFlow code and pre-trained models for BERT☆38,939Updated 8 months ago
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Auto…☆13,509Updated this week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.☆37,675Updated this week
- State-of-the-Art Text Embeddings☆16,347Updated this week
- A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)☆5,554Updated this week
- XLNet: Generalized Autoregressive Pretraining for Language Understanding☆6,185Updated last year
- Code for the paper "Language Models are Unsupervised Multitask Learners"☆23,251Updated 7 months ago
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"☆11,614Updated 3 months ago
- Fast and memory-efficient exact attention☆16,664Updated this week
- Pretrain, finetune ANY AI model of ANY size on multiple GPUs, TPUs with zero code changes.☆29,216Updated this week
- A framework for training and evaluating AI models on a variety of openly available dialogue datasets.☆10,517Updated last year
- GPT-3: Language Models are Few-Shot Learners☆15,753Updated 4 years ago
- Repo for external large-scale work☆6,523Updated 11 months ago
- Open Source Neural Machine Translation and (Large) Language Models in PyTorch☆6,855Updated 3 weeks ago
- A PyTorch implementation of the Transformer model in "Attention is All You Need".☆9,103Updated 11 months ago
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training☆21,664Updated 7 months ago
- CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image☆28,247Updated 8 months ago
- Google Research☆35,222Updated last week
- An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.☆38,262Updated last week
- An annotated implementation of the Transformer paper.☆6,132Updated 11 months ago