lucidrains / discrete-key-value-bottleneck-pytorchLinks
Implementation of Discrete Key / Value Bottleneck, in Pytorch
☆88Updated 2 years ago
Alternatives and similar repositories for discrete-key-value-bottleneck-pytorch
Users that are interested in discrete-key-value-bottleneck-pytorch are comparing it to the libraries listed below
Sorting:
- Implementation of Gated State Spaces, from the paper "Long Range Language Modeling via Gated State Spaces", in Pytorch☆101Updated 2 years ago
- Implementation of "compositional attention" from MILA, a multi-head attention variant that is reframed as a two-step attention process wi…☆51Updated 3 years ago
- Implementation of Hourglass Transformer, in Pytorch, from Google and OpenAI☆91Updated 3 years ago
- Implementation of Mega, the Single-head Attention with Multi-headed EMA architecture that currently holds SOTA on Long Range Arena☆204Updated last year
- ☆51Updated last year
- JAX implementation ViT-VQGAN☆83Updated 2 years ago
- Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts☆120Updated 9 months ago
- Standalone Product Key Memory module in Pytorch - for augmenting Transformer models☆82Updated last year
- Implementation of some personal helper functions for Einops, my most favorite tensor manipulation library ❤️☆55Updated 2 years ago
- Implementation of Retrieval-Augmented Denoising Diffusion Probabilistic Models in Pytorch☆65Updated 3 years ago
- ☆74Updated 2 years ago
- codebase for the SIMAT dataset and evaluation☆38Updated 3 years ago
- Un-*** 50 billions multimodality dataset☆23Updated 2 years ago
- Implementation of 🌻 Mirasol, SOTA Multimodal Autoregressive model out of Google Deepmind, in Pytorch☆89Updated last year
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixing☆50Updated 3 years ago
- Implementation of a Transformer that Ponders, using the scheme from the PonderNet paper☆81Updated 3 years ago
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆44Updated 2 years ago
- Implementation of the Kalman Filtering Attention proposed in "Kalman Filtering Attention for User Behavior Modeling in CTR Prediction"☆58Updated last year
- The official repository for our paper "The Dual Form of Neural Networks Revisited: Connecting Test Time Predictions to Training Patterns …☆16Updated last month
- Official Repository of Pretraining Without Attention (BiGS), BiGS is the first model to achieve BERT-level transfer learning on the GLUE …☆114Updated last year
- Exploration into the proposed "Self Reasoning Tokens" by Felipe Bonetto☆56Updated last year
- Implementation of Zorro, Masked Multimodal Transformer, in Pytorch☆97Updated last year
- Implementation of MaMMUT, a simple vision-encoder text-decoder architecture for multimodal tasks from Google, in Pytorch☆103Updated last year
- TF/Keras code for DiffStride, a pooling layer with learnable strides.☆124Updated 3 years ago
- Implementation of Recurrent Interface Network (RIN), for highly efficient generation of images and video without cascading networks, in P…☆205Updated last year
- Code for ICLR 2023 Paper, "Stable Target Field for Reduced Variance Score Estimation in Diffusion Models”☆76Updated 2 years ago
- Implementation of an Attention layer where each head can attend to more than just one token, using coordinate descent to pick topk☆46Updated 2 years ago
- A GPT, made only of MLPs, in Jax☆58Updated 4 years ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 3 years ago
- Another attempt at a long-context / efficient transformer by me☆38Updated 3 years ago