lilianweng / transformer-tensorflow
Implementation of Transformer Model in Tensorflow
☆456Updated last year
Related projects ⓘ
Alternatives and complementary repositories for transformer-tensorflow
- A Keras+TensorFlow Implementation of the Transformer: Attention Is All You Need☆709Updated 3 years ago
- Attention mechanism for processing sequential data that considers the context for each timestamp.☆657Updated 2 years ago
- Implementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with Ten…☆347Updated 9 months ago
- Minimal Seq2Seq model with Attention for Neural Machine Translation in PyTorch☆691Updated 3 years ago
- TensorFlow implementation of 'Attention Is All You Need (2017. 6)'☆350Updated 6 years ago
- A wrapper layer for stacking layers horizontally☆227Updated 2 years ago
- Keras library for building (Universal) Transformers, facilitating BERT and GPT models☆533Updated 4 years ago
- Sequence-to-sequence model with LSTM encoder/decoders and attention☆1,261Updated 3 years ago
- Sequence to Sequence Models with PyTorch☆736Updated 2 years ago
- Deep Reinforcement Learning For Sequence to Sequence Models☆765Updated last year
- Sequence-to-Sequence learning using PyTorch☆523Updated 5 years ago
- Visualizing RNNs using the attention mechanism☆748Updated 5 years ago
- An open source framework for seq2seq models in PyTorch.☆1,498Updated last year
- Transformer implemented in Keras☆371Updated 2 years ago
- Hierarchical Attention Networks for Document Classification in PyTorch☆606Updated 4 years ago
- Transformer training code for sequential tasks☆609Updated 3 years ago
- Code for the paper "Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks"☆578Updated 5 years ago
- LSTM and QRNN Language Model Toolkit for PyTorch☆1,960Updated 2 years ago
- A repository containing tutorials for practical NLP using PyTorch☆530Updated 5 years ago
- Keras implementation of BERT with pre-trained weights☆813Updated 5 years ago
- An LSTM in PyTorch with best practices (weight dropout, forget bias, etc.) built-in. Fully compatible with PyTorch LSTM.☆133Updated 4 years ago
- Neural Turing Machines (NTM) - PyTorch Implementation☆590Updated 6 years ago
- Implementation of Universal Transformer in Pytorch☆258Updated 6 years ago
- Single Headed Attention RNN - "Stop thinking with your head"☆1,179Updated 2 years ago
- Dynamic seq2seq in TensorFlow, step by step☆996Updated 7 years ago
- Keras Layer implementation of Attention for Sequential models☆445Updated last year
- Pervasive Attention: 2D Convolutional Networks for Sequence-to-Sequence Prediction☆502Updated 3 years ago
- Sequence to sequence learning using TensorFlow.☆392Updated 6 years ago
- Attention-based sequence to sequence learning☆389Updated 5 years ago
- [ICLR'19] Trellis Networks for Sequence Modeling☆473Updated 5 years ago