harvardnlp / annotated-transformerLinks
An annotated implementation of the Transformer paper.
☆6,354Updated last year
Alternatives and similar repositories for annotated-transformer
Users that are interested in annotated-transformer are comparing it to the libraries listed below
Sorting:
- A PyTorch implementation of the Transformer model in "Attention is All You Need".☆9,283Updated last year
- Google AI 2018 BERT pytorch implementation☆6,433Updated last year
- Transformer: PyTorch Implementation of "Attention Is All You Need"☆3,861Updated 11 months ago
- BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)☆7,535Updated last month
- A TensorFlow Implementation of the Transformer: Attention Is All You Need☆4,382Updated 2 years ago
- ☆3,661Updated 2 years ago
- Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.☆16,290Updated 2 years ago
- Tutorials on implementing a few sequence-to-sequence (seq2seq) models with PyTorch and TorchText.☆5,579Updated last year
- Unsupervised text tokenizer for Neural Network-based text generation.☆11,087Updated this week
- Transformer seq2seq model, program that can build a language translator from parallel corpus☆1,401Updated 2 years ago
- Must-read Papers on pre-trained language models.☆3,358Updated 2 years ago
- XLNet: Generalized Autoregressive Pretraining for Language Understanding☆6,185Updated 2 years ago
- A concise but complete full-attention transformer with a set of promising experimental features from various papers☆5,441Updated last week
- Open Source Neural Machine Translation and (Large) Language Models in PyTorch☆6,906Updated 4 months ago
- Longformer: The Long-Document Transformer☆2,147Updated 2 years ago
- ☆11,564Updated 4 months ago
- Code and model for the paper "Improving Language Understanding by Generative Pre-Training"☆2,222Updated 6 years ago
- Simple transformer implementation from scratch in pytorch. (archival, latest version on codeberg)☆1,087Updated 3 months ago
- Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"☆6,387Updated 2 months ago
- Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.☆3,113Updated last year
- Models, data loaders and abstractions for language processing, powered by PyTorch☆3,546Updated last week
- [EMNLP 2021] SimCSE: Simple Contrastive Learning of Sentence Embeddings https://arxiv.org/abs/2104.08821☆3,576Updated 9 months ago
- Unsupervised Word Segmentation for Neural Machine Translation and Text Generation☆2,242Updated 11 months ago
- Ongoing research training transformer models at scale☆12,835Updated this week
- Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Py…☆23,353Updated 4 months ago
- Reading list for research topics in multimodal machine learning☆6,540Updated 10 months ago
- 🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.☆19,028Updated last week
- 🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (i…☆8,941Updated this week
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities☆21,520Updated 2 weeks ago
- Fast and memory-efficient exact attention☆18,340Updated this week