RMichaelSwan / MogrifierLSTMLinks
A quick walk-through of the innards of LSTMs and a naive implementation of the Mogrifier LSTM paper in PyTorch
☆78Updated 5 years ago
Alternatives and similar repositories for MogrifierLSTM
Users that are interested in MogrifierLSTM are comparing it to the libraries listed below
Sorting:
- Pytorch implementation of "Block Recurrent Transformers" (Hutchins & Schlag et al., 2022)☆84Updated 3 years ago
- ☆84Updated 5 years ago
- A simple module consistently outperforms self-attention and Transformer model on main NMT datasets with SoTA performance.☆86Updated 2 years ago
- Implementing SYNTHESIZER: Rethinking Self-Attention in Transformer Models using Pytorch☆70Updated 5 years ago
- Implementation of Mogrifier LSTM in PyTorch☆34Updated 5 years ago
- [EMNLP'19] Summary for Transformer Understanding☆53Updated 5 years ago
- A PyTorch implementation of the paper - "Synthesizer: Rethinking Self-Attention in Transformer Models"☆73Updated 2 years ago
- Learning to Encode Position for Transformer with Continuous Dynamical Model☆59Updated 5 years ago
- Multi-head attention in PyTorch☆154Updated 6 years ago
- ☆64Updated 4 years ago
- Code for Multi-Head Attention: Collaborate Instead of Concatenate☆151Updated 2 years ago
- LAnguage Modelling Benchmarks☆138Updated 5 years ago
- Code for the ACL2020 paper Character-Level Translation with Self-Attention☆31Updated 5 years ago
- ECML 2019: Graph Neural Networks for Multi-Label Classification☆90Updated last year
- Two-Layer Hierarchical Softmax Implementation for PyTorch☆70Updated 4 years ago
- [ICML 2020] code for the flooding regularizer proposed in "Do We Need Zero Training Loss After Achieving Zero Training Error?"☆93Updated 2 years ago
- Code for "Understanding and Improving Layer Normalization"☆46Updated 5 years ago
- ☆20Updated 5 years ago
- Implementation of RealFormer using pytorch☆101Updated 4 years ago
- Variational Transformers for Diverse Response Generation☆81Updated last year
- This repository contain various types of attention mechanism like Bahdanau , Soft attention , Additive Attention , Hierarchical Attention…☆125Updated 4 years ago
- This in my Demo of Chen et al. "GradNorm: Gradient Normalization for Adaptive Loss Balancing in Deep Multitask Networks" ICML 2018☆181Updated 3 years ago
- 基于Transformer的单模型、多尺度的VAE模型☆57Updated 4 years ago
- a simple pytorch implement of Multi-Sample Dropout☆57Updated 6 years ago
- pytorch neural network attention mechanism☆147Updated 6 years ago
- Code for the paper "Adaptive Transformers for Learning Multimodal Representations" (ACL SRW 2020)☆43Updated 3 years ago
- Implements Reformer: The Efficient Transformer in pytorch.☆86Updated 5 years ago
- PyTorch Implementation of the Multi-gate Mixture-of-Experts with Exclusivity (MMoEEx)☆32Updated 4 years ago
- code for Explicit Sparse Transformer☆61Updated 2 years ago
- The code of Encoding Word Order in Complex-valued Embedding☆42Updated 6 years ago