aliutkus / spe
Relative Positional Encoding for Transformers with Linear Complexity
☆62Updated 3 years ago
Alternatives and similar repositories for spe:
Users that are interested in spe are comparing it to the libraries listed below
- Code for the paper PermuteFormer☆42Updated 3 years ago
- Sequence Modeling with Structured State Spaces☆63Updated 2 years ago
- Axial Positional Embedding for Pytorch☆77Updated last month
- Jax/Flax implementation of Variational-DiffWave.☆40Updated 3 years ago
- Implementation of Gated State Spaces, from the paper "Long Range Language Modeling via Gated State Spaces", in Pytorch☆99Updated 2 years ago
- ☆74Updated 4 years ago
- Implements Reformer: The Efficient Transformer in pytorch.☆85Updated 5 years ago
- A variant of Transformer-XL where the memory is updated not with a queue, but with attention☆48Updated 4 years ago
- [EMNLP'19] Summary for Transformer Understanding☆53Updated 5 years ago
- Implementation of N-Grammer, augmenting Transformers with latent n-grams, in Pytorch☆73Updated 2 years ago
- Codes for paper <InteL-VAEs: Adding Inductive Biases to VariationalAuto-Encoders via Intermediary Latents>.☆19Updated 3 years ago
- Code for the ICML'20 paper "Improving Transformer Optimization Through Better Initialization"☆88Updated 4 years ago
- Continuous Augmented Positional Embeddings (CAPE) implementation for PyTorch☆40Updated 2 years ago
- Skyformer: Remodel Self-Attention with Gaussian Kernel and Nystr\"om Method (NeurIPS 2021)☆60Updated 2 years ago
- Implementation of Long-Short Transformer, combining local and global inductive biases for attention over long sequences, in Pytorch☆118Updated 3 years ago
- Tensorflow Implementation of "Theory and Experiments on Vector Quantized Autoencoders"☆14Updated 6 years ago
- Code for ICLR 2021 Paper, "Anytime Sampling for Autoregressive Models via Ordered Autoencoding"☆26Updated last year
- Implementation of Insertion-deletion Denoising Diffusion Probabilistic Models☆30Updated 2 years ago
- PyTorch implementation of the paper "NanoFlow: Scalable Normalizing Flows with Sublinear Parameter Complexity." (NeurIPS 2020)☆65Updated 4 years ago
- Representation learning for NLP @ JSALT19☆38Updated 4 years ago
- Cascaded Text Generation with Markov Transformers☆129Updated 2 years ago
- Code for "Understanding and Improving Layer Normalization"☆46Updated 5 years ago
- Implementation of the Kalman Filtering Attention proposed in "Kalman Filtering Attention for User Behavior Modeling in CTR Prediction"☆57Updated last year
- Official code repository of the paper Linear Transformers Are Secretly Fast Weight Programmers.☆104Updated 3 years ago
- Official code repository of the paper Learning Associative Inference Using Fast Weight Memory by Schlag et al.☆28Updated 4 years ago
- On Variational Learning of Controllable Representations for Text without Supervision https://arxiv.org/abs/1905.11975☆27Updated 4 years ago
- Implementation of Memory-Compressed Attention, from the paper "Generating Wikipedia By Summarizing Long Sequences"☆70Updated 2 years ago
- Implementation of Perceiver AR, Deepmind's new long-context attention network based on Perceiver architecture, in Pytorch☆87Updated 2 years ago
- Implementation of the BASIS algorithm for source separation with deep generative priors☆39Updated 2 years ago
- Implementation of Fast Transformer in Pytorch☆173Updated 3 years ago