jayparks / transformer
A Pytorch Implementation of "Attention is All You Need" and "Weighted Transformer Network for Machine Translation"
☆553Updated 4 years ago
Alternatives and similar repositories for transformer:
Users that are interested in transformer are comparing it to the libraries listed below
- pytorch implementation of Attention is all you need☆238Updated 3 years ago
- Minimal Seq2Seq model with Attention for Neural Machine Translation in PyTorch☆696Updated 4 years ago
- ☆875Updated 9 months ago
- Transformer seq2seq model, program that can build a language translator from parallel corpus☆1,383Updated last year
- ☆3,639Updated 2 years ago
- [ICLR 2020] Lite Transformer with Long-Short Range Attention☆605Updated 8 months ago
- PyTorch implementation of batched bi-RNN encoder and attention-decoder.☆279Updated 6 years ago
- Pytorch Implementation of Google BERT☆593Updated 4 years ago
- PyTorch implementation of beam search decoding for seq2seq models☆337Updated 2 years ago
- Code for the paper "Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks"☆578Updated 5 years ago
- Reformer, the efficient Transformer, in Pytorch☆2,151Updated last year
- Pytorch implementation of R-Transformer. Some parts of the code are adapted from the implementation of TCN and Transformer.☆227Updated 5 years ago
- Minimal RNN classifier with self-attention in Pytorch☆150Updated 3 years ago
- My take on a practical implementation of Linformer for Pytorch.☆412Updated 2 years ago
- ☆166Updated 5 years ago
- Transformer implementation in PyTorch.☆478Updated 6 years ago
- some attention implements☆1,442Updated 5 years ago
- A simplified PyTorch implementation of "SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient." (Yu, Lantao, et al.)☆645Updated 6 years ago
- DeLighT: Very Deep and Light-Weight Transformers☆468Updated 4 years ago
- A implementation of SeqGAN in PyTorch, following the implementation in tensorflow.☆264Updated 6 years ago
- PyTorch Re-Implementation of "Generating Sentences from a Continuous Space" by Bowman et al 2015 https://arxiv.org/abs/1511.06349☆590Updated 2 months ago
- Sequence to Sequence Models with PyTorch☆737Updated 2 years ago
- Examples of using sparse attention, as in "Generating Long Sequences with Sparse Transformers"☆1,559Updated 4 years ago
- Natural Language Processing Tutorial for Deep Learning Researchers☆1,113Updated 2 years ago
- Sequence Generation Model for Multi-label Classification (COLING 2018)☆437Updated 5 years ago
- list of efficient attention modules☆995Updated 3 years ago
- Implementation of Universal Transformer in Pytorch☆259Updated 6 years ago
- ☆267Updated 2 years ago
- ☆230Updated 5 years ago
- pytorch implementation of "Get To The Point: Summarization with Pointer-Generator Networks"☆910Updated 2 years ago