10-zin / Synthesizer
A PyTorch implementation of the paper - "Synthesizer: Rethinking Self-Attention in Transformer Models"
☆72Updated 2 years ago
Alternatives and similar repositories for Synthesizer:
Users that are interested in Synthesizer are comparing it to the libraries listed below
- Code for "Understanding and Improving Layer Normalization"☆46Updated 5 years ago
- Implementation of COCO-LM, Correcting and Contrasting Text Sequences for Language Model Pretraining, in Pytorch☆45Updated 4 years ago
- Implementation of the retriever distillation procedure as outlined in the paper "Distilling Knowledge from Reader to Retriever"☆32Updated 4 years ago
- Code for EMNLP 2020 paper CoDIR☆41Updated 2 years ago
- 基于Transformer的单模型、多尺度的VAE模型☆55Updated 3 years ago
- Implementation of Mixout with PyTorch☆74Updated 2 years ago
- DisCo Transformer for Non-autoregressive MT☆78Updated 2 years ago
- The implementation of multi-branch attentive Transformer (MAT).☆33Updated 4 years ago
- Cascaded Text Generation with Markov Transformers☆129Updated last year
- Implementing SYNTHESIZER: Rethinking Self-Attention in Transformer Models using Pytorch☆70Updated 4 years ago
- ICLR2019, Multilingual Neural Machine Translation with Knowledge Distillation☆70Updated 4 years ago
- ☆63Updated 2 years ago
- ☆20Updated 5 years ago
- ☆22Updated 3 years ago
- Implements Reformer: The Efficient Transformer in pytorch.☆84Updated 5 years ago
- ☆32Updated 3 years ago
- Repository containing code for the paper "Meta-Learning with Sparse Experience Replay for Lifelong Language Learning".☆21Updated last year
- This repository contains the code for running the character-level Sandwich Transformers from our ACL 2020 paper on Improving Transformer …☆55Updated 4 years ago
- PyTorch implementation of Pay Attention to MLPs☆40Updated 3 years ago
- This repo provides the code for the ACL 2020 paper "Evidence-Aware Inferential Text Generation with Vector Quantised Variational AutoEnco…☆53Updated 4 years ago
- Code for the paper "Query-Key Normalization for Transformers"☆37Updated 3 years ago
- ☆51Updated 4 years ago
- Implementation of Long-Short Transformer, combining local and global inductive biases for attention over long sequences, in Pytorch☆118Updated 3 years ago
- A simple module consistently outperforms self-attention and Transformer model on main NMT datasets with SoTA performance.☆86Updated last year
- Code for Multi-Head Attention: Collaborate Instead of Concatenate☆152Updated last year
- ☆13Updated 5 years ago
- Code for ACL2020 "Jointly Masked Sequence-to-Sequence Model for Non-Autoregressive Neural Machine Translation"☆39Updated 4 years ago
- Code and data to accompany the camera-ready version of "Cross-Attention is All You Need: Adapting Pretrained Transformers for Machine Tra…☆29Updated 3 years ago
- Official PyTorch implementation of Time-aware Large Kernel (TaLK) Convolutions (ICML 2020)☆29Updated 4 years ago
- Implementation of RealFormer using pytorch☆101Updated 4 years ago