zimmerrol / attention-is-all-you-need-keras
Implementation of the Transformer architecture described by Vaswani et al. in "Attention Is All You Need"
☆28Updated 5 years ago
Related projects ⓘ
Alternatives and complementary repositories for attention-is-all-you-need-keras
- An Attention Layer in Keras☆43Updated 5 years ago
- Collection of custom layers and utility functions for Keras which are missing in the main framework.☆62Updated 4 years ago
- Layer normalization implemented in Keras☆60Updated 2 years ago
- Keras implementation of “Gated Linear Unit ”☆23Updated 7 months ago
- Position embedding layers in Keras☆59Updated 2 years ago
- attention block for keras Functional Model with only tensorflow backend☆26Updated 5 years ago
- Transformer-XL with checkpoint loader☆68Updated 2 years ago
- Tensorflow Implementation of Densely Connected Bidirectional LSTM with Applications to Sentence Classification☆48Updated 6 years ago
- My implementation of "Hierarchical Attention Networks for Document Classification" in Keras☆26Updated 6 years ago
- 自注意力与文本分类☆119Updated 6 years ago
- Simple Tensorflow Implementation of "A Structured Self-attentive Sentence Embedding" (ICLR 2017)☆92Updated 6 years ago
- CapsNet for NLP☆68Updated 5 years ago
- Tensorflow implementation of Semi-supervised Sequence Learning (https://arxiv.org/abs/1511.01432)☆82Updated 2 years ago
- seq2seq attention in keras☆40Updated 5 years ago
- Sequence to Sequence and attention from scratch using Tensorflow☆29Updated 7 years ago
- Ordered Neurons LSTM☆30Updated 2 years ago
- Implement en-fr translation task by implenting seq2seq, encoder-decoder in RNN layers with Attention mechanism and Beamsearch inference d…☆21Updated 6 years ago
- Tensorflow implementation of "A Structured Self-Attentive Sentence Embedding"☆194Updated 3 years ago
- Tensorflow implementation of the Convolutional Recurrent Neural Network model with max pooling and attentive pooling, for relation classi…☆48Updated 6 years ago
- Efficient Transformers for research, PyTorch and Tensorflow using Locality Sensitive Hashing☆93Updated 4 years ago
- Implementation of Hierarchical Attention Networks as presented in https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf☆58Updated 6 years ago
- Multilingual hierarchical attention networks toolkit☆78Updated 4 years ago
- Multi-Task Learning in NLP☆95Updated 6 years ago
- Reproducing Character-Level-Language-Modeling with Deeper Self-Attention in PyTorch☆61Updated 5 years ago
- This repository contain various types of attention mechanism like Bahdanau , Soft attention , Additive Attention , Hierarchical Attention…☆122Updated 3 years ago
- Tensorflow Implementation of Variational Attention for Sequence to Sequence Models (COLING 2018)☆69Updated 4 years ago
- Re-implementation of ELMo on Keras☆135Updated last year
- ☆39Updated 7 years ago
- Keras implementation of Nested LSTMs☆90Updated 5 years ago
- An extension on the Hierachical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion, our implementation is in Tensorf…☆71Updated 7 years ago