phohenecker / pytorch-transformerLinks
A PyTorch implementation of the Transformer model from "Attention Is All You Need".
☆59Updated 6 years ago
Alternatives and similar repositories for pytorch-transformer
Users that are interested in pytorch-transformer are comparing it to the libraries listed below
Sorting:
- The Annotated Encoder Decoder with Attention☆167Updated 4 years ago
- Cascaded Text Generation with Markov Transformers☆129Updated 2 years ago
- ☆121Updated 6 years ago
- Highway network implemented in pytorch☆80Updated 8 years ago
- Code examples for CMU CS11-731, Machine Translation and Sequence-to-sequence Models☆35Updated 6 years ago
- Code for the Eager Translation Model from the paper You May Not Need Attention☆295Updated 6 years ago
- Training Transformer-XL on 128 GPUs☆141Updated 5 years ago
- Adaptive Softmax implementation for PyTorch☆81Updated 6 years ago
- Code for EMNLP18 paper "Spherical Latent Spaces for Stable Variational Autoencoders"☆170Updated 6 years ago
- LaNMT: Latent-variable Non-autoregressive Neural Machine Translation with Deterministic Inference☆79Updated 4 years ago
- ☆47Updated 6 years ago
- Checking the interpretability of attention on text classification models☆49Updated 6 years ago
- Reproducing Character-Level-Language-Modeling with Deeper Self-Attention in PyTorch☆62Updated 6 years ago
- Generative Flow based Sequence-to-Sequence Toolkit written in Python.☆246Updated 5 years ago
- a Pytorch implementation of the Reformer Network (https://openreview.net/pdf?id=rkgNKkHtvB)☆53Updated 3 years ago
- Code for "Language GANs Falling Short"☆59Updated 4 years ago
- An LSTM in PyTorch with best practices (weight dropout, forget bias, etc.) built-in. Fully compatible with PyTorch LSTM.☆134Updated 5 years ago
- ☆65Updated 5 years ago
- ☆178Updated 5 years ago
- Code inspired by Unsupervised Machine Translation Using Monolingual Corpora Only☆50Updated last year
- Embedding Quantization (Compress Word Embeddings)☆85Updated 6 years ago
- PyTorch DataLoader for seq2seq☆85Updated 6 years ago
- Factorization of the neural parameter space for zero-shot multi-lingual and multi-task transfer☆39Updated 5 years ago
- ☆119Updated 6 years ago
- Code for Multi-Head Attention: Collaborate Instead of Concatenate☆151Updated 2 years ago
- Sparse and structured neural attention mechanisms☆225Updated 5 years ago
- SparseMAP: differentiable sparse structure inference☆112Updated 6 years ago
- Latent Alignment and Variational Attention☆327Updated 7 years ago
- Recurrent Variational Autoencoder with Dilated Convolutions that generates sequential data implemented in pytorch☆71Updated 4 years ago
- Pytorch and Torchtext implementation of Sequence to sequence☆59Updated 7 years ago