harvardnlp / annotated-transformerLinks
An annotated implementation of the Transformer paper.
☆6,885Updated last year
Alternatives and similar repositories for annotated-transformer
Users that are interested in annotated-transformer are comparing it to the libraries listed below
Sorting:
- A PyTorch implementation of the Transformer model in "Attention is All You Need".☆9,576Updated last year
- Google AI 2018 BERT pytorch implementation☆6,510Updated 2 years ago
- BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)☆7,849Updated 7 months ago
- Transformer: PyTorch Implementation of "Attention Is All You Need"☆4,356Updated 5 months ago
- ☆3,682Updated 3 years ago
- A TensorFlow Implementation of the Transformer: Attention Is All You Need☆4,456Updated 2 years ago
- Tutorials on implementing a few sequence-to-sequence (seq2seq) models with PyTorch and TorchText.☆5,664Updated last year
- Transformer seq2seq model, program that can build a language translator from parallel corpus☆1,422Updated 2 years ago
- Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.☆16,854Updated 2 years ago
- A concise but complete full-attention transformer with a set of promising experimental features from various papers☆5,743Updated last week
- The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights --…☆36,114Updated this week
- A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch☆8,885Updated last week
- Reading list for research topics in multimodal machine learning☆6,771Updated last year
- ☆12,161Updated 9 months ago
- Natural Language Processing Tutorial for Deep Learning Researchers☆14,826Updated last year
- Must-read Papers on pre-trained language models.☆3,367Updated 3 years ago
- 🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (i…☆9,411Updated 2 weeks ago
- Code and model for the paper "Improving Language Understanding by Generative Pre-Training"☆2,263Updated 6 years ago
- Open Source Neural Machine Translation and (Large) Language Models in PyTorch☆6,981Updated 2 months ago
- XLNet: Generalized Autoregressive Pretraining for Language Understanding☆6,177Updated 2 years ago
- [EMNLP 2021] SimCSE: Simple Contrastive Learning of Sentence Embeddings https://arxiv.org/abs/2104.08821☆3,629Updated last year
- Longformer: The Long-Document Transformer☆2,178Updated 2 years ago
- PyTorch implementation of MoCo: https://arxiv.org/abs/1911.05722☆5,106Updated 3 months ago
- Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Py…☆24,737Updated last week
- 🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.☆20,347Updated 2 weeks ago
- Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)☆9,334Updated last month
- AI conference deadline countdowns☆5,970Updated last year
- Ongoing research training transformer models at scale☆14,758Updated this week
- Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"☆6,468Updated last month
- A small package to create visualizations of PyTorch execution graphs☆3,481Updated last year