jayparks / transformer
A Pytorch Implementation of "Attention is All You Need" and "Weighted Transformer Network for Machine Translation"
☆556Updated 4 years ago
Alternatives and similar repositories for transformer:
Users that are interested in transformer are comparing it to the libraries listed below
- Minimal Seq2Seq model with Attention for Neural Machine Translation in PyTorch☆699Updated 4 years ago
- pytorch implementation of Attention is all you need☆238Updated 3 years ago
- [ICLR 2020] Lite Transformer with Long-Short Range Attention☆607Updated 9 months ago
- Transformer seq2seq model, program that can build a language translator from parallel corpus☆1,395Updated last year
- PyTorch implementation of batched bi-RNN encoder and attention-decoder.☆279Updated 6 years ago
- PyTorch implementation of beam search decoding for seq2seq models☆337Updated 2 years ago
- some attention implements☆1,445Updated 5 years ago
- list of efficient attention modules☆1,003Updated 3 years ago
- Reformer, the efficient Transformer, in Pytorch☆2,164Updated last year
- Code for the paper "Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks"☆579Updated 5 years ago
- Transformer implementation in PyTorch.☆486Updated 6 years ago
- ☆876Updated 11 months ago
- My take on a practical implementation of Linformer for Pytorch.☆415Updated 2 years ago
- Pytorch Implementation of Google BERT☆593Updated 5 years ago
- ☆3,645Updated 2 years ago
- Hierarchical Attention Networks for Document Classification in PyTorch☆603Updated 5 years ago
- A simplified PyTorch implementation of "SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient." (Yu, Lantao, et al.)☆646Updated 6 years ago
- Multi-head attention in PyTorch☆152Updated 6 years ago
- ☆166Updated 6 years ago
- A implementation of SeqGAN in PyTorch, following the implementation in tensorflow.☆264Updated 6 years ago
- Sequence to Sequence Models with PyTorch☆738Updated 3 years ago
- ☆231Updated 5 years ago
- Examples of using sparse attention, as in "Generating Long Sequences with Sparse Transformers"☆1,569Updated 4 years ago
- Pytorch implementation of R-Transformer. Some parts of the code are adapted from the implementation of TCN and Transformer.☆228Updated 5 years ago
- (Linear-chain) Conditional random field in PyTorch.☆960Updated 10 months ago
- An open source framework for seq2seq models in PyTorch.☆1,507Updated 2 years ago
- Sequence Generation Model for Multi-label Classification (COLING 2018)☆437Updated 5 years ago
- Transformer with Untied Positional Encoding (TUPE). Code of paper "Rethinking Positional Encoding in Language Pre-training". Improve exis…☆251Updated 3 years ago
- A Structured Self-attentive Sentence Embedding☆492Updated 5 years ago
- torchtext使用总结,从零开始逐步实现了torchtext文本预处理过程,包括截断补长,词表构建,使用预训练词向量,构建可用于PyTorch的可迭代数据等步骤。并结合Pytorch实现LSTM.☆173Updated 5 years ago