google-deepmind / lambLinks
LAnguage Modelling Benchmarks
☆138Updated 5 years ago
Alternatives and similar repositories for lamb
Users that are interested in lamb are comparing it to the libraries listed below
Sorting:
- A simple module consistently outperforms self-attention and Transformer model on main NMT datasets with SoTA performance.☆86Updated 2 years ago
- ☆47Updated 6 years ago
- PyTorch DataLoader for seq2seq☆85Updated 6 years ago
- Cascaded Text Generation with Markov Transformers☆129Updated 2 years ago
- An LSTM in PyTorch with best practices (weight dropout, forget bias, etc.) built-in. Fully compatible with PyTorch LSTM.☆134Updated 5 years ago
- Pytorch implementation of R-Transformer. Some parts of the code are adapted from the implementation of TCN and Transformer.☆230Updated 6 years ago
- ☆178Updated 5 years ago
- ☆219Updated 5 years ago
- NeurIPS 2019 - Learning Data Manipulation for Augmentation and Weighting☆110Updated 5 years ago
- ☆65Updated 5 years ago
- PyTorch implementations of LSTM Variants (Dropout + Layer Norm)☆137Updated 4 years ago
- A smoother activation function (undergrad code)☆115Updated 5 years ago
- Code for EMNLP 2019 paper "Attention is not not Explanation"☆58Updated 4 years ago
- Implementation of Universal Transformer in Pytorch☆265Updated 7 years ago
- Code for reproducing experiments in our ACL 2019 paper "Probing Neural Network Comprehension of Natural Language Arguments"☆54Updated 3 years ago
- A PyTorch implementation of the Transformer model from "Attention Is All You Need".☆60Updated 6 years ago
- PyTorch code for meta seq2seq learning☆43Updated 5 years ago
- The Annotated Encoder Decoder with Attention☆167Updated 4 years ago
- Code for EMNLP18 paper "Spherical Latent Spaces for Stable Variational Autoencoders"☆171Updated 6 years ago
- Checking the interpretability of attention on text classification models☆49Updated 6 years ago
- Code for "Language GANs Falling Short"☆59Updated 4 years ago
- Boolean Question Answering with multi-task learning and uses large LM embeddings like BERT, RoBERTa☆18Updated 6 years ago
- Bayesian Deep Active Learning for Natural Language Processing Tasks☆147Updated 7 years ago
- Two-Layer Hierarchical Softmax Implementation for PyTorch☆70Updated 4 years ago
- Variational Attention for Sequence to Sequence Models☆20Updated 7 years ago
- Non-Monotonic Sequential Text Generation (ICML 2019)☆72Updated 6 years ago
- Factorization of the neural parameter space for zero-shot multi-lingual and multi-task transfer☆39Updated 5 years ago
- Code for Multi-Head Attention: Collaborate Instead of Concatenate☆153Updated 2 years ago
- a Pytorch implementation of the Reformer Network (https://openreview.net/pdf?id=rkgNKkHtvB)☆53Updated 3 years ago
- Code for "Strong Baselines for Neural Semi-supervised Learning under Domain Shift" (Ruder & Plank, 2018 ACL)☆61Updated 2 years ago