minhtannguyen / transformer-mgk
This is the public github for our paper "Transformer with a Mixture of Gaussian Keys"
☆26Updated 2 years ago
Alternatives and similar repositories for transformer-mgk:
Users that are interested in transformer-mgk are comparing it to the libraries listed below
- Code for the PAPA paper☆27Updated 2 years ago
- Skyformer: Remodel Self-Attention with Gaussian Kernel and Nystr\"om Method (NeurIPS 2021)☆60Updated 2 years ago
- Implementation of "compositional attention" from MILA, a multi-head attention variant that is reframed as a two-step attention process wi…☆50Updated 2 years ago
- Implementation of Cross Transformer for spatially-aware few-shot transfer, in Pytorch☆52Updated 4 years ago
- Official PyTorch implementation for NeurIPS'24 paper "Knowledge Composition using Task Vectors with Learned Anisotropic Scaling"☆19Updated last month
- The official repository for our paper "The Dual Form of Neural Networks Revisited: Connecting Test Time Predictions to Training Patterns …☆16Updated last year
- [ICLR 2023] “ Layer Grafted Pre-training: Bridging Contrastive Learning And Masked Image Modeling For Better Representations”, Ziyu Jian…☆24Updated 2 years ago
- Distributional Sliced-Wasserstein distance code☆49Updated 8 months ago
- Diffusion based transformer, in PyTorch (Experimental).☆24Updated 2 years ago
- Curse-of-memory phenomenon of RNNs in sequence modelling☆19Updated last week
- ☆21Updated 2 years ago
- Github code for the paper Maximum Class Separation as Inductive Bias in One Matrix. Arxiv link: https://arxiv.org/abs/2206.08704☆29Updated last year
- ☆26Updated 3 years ago
- ☆22Updated last year
- ImageNet-12k subset of ImageNet-21k (fall11)☆21Updated last year
- Position Prediction as an Effective Pretraining Strategy☆8Updated 2 years ago
- Mixture of Attention Heads☆43Updated 2 years ago
- DiWA: Diverse Weight Averaging for Out-of-Distribution Generalization☆29Updated 2 years ago
- Bag of MLP☆20Updated 3 years ago
- Energy-Based Models for Continual Learning Official Repository (PyTorch)☆41Updated 2 years ago
- Self-Distillation with weighted ground-truth targets; ResNet and Kernel Ridge Regression☆17Updated 3 years ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 2 years ago
- Code for paper "Can contrastive learning avoid shortcut solutions?" NeurIPS 2021.☆47Updated 3 years ago
- Implementation of TableFormer, Robust Transformer Modeling for Table-Text Encoding, in Pytorch☆37Updated 3 years ago
- ☆29Updated 2 years ago
- Official implementation of the paper "Provable Stochastic Optimization for Global Contrastive Learning: Small Batch Does Not Harm Perform…☆20Updated 2 years ago
- [NeurIPS 2022 Spotlight] This is the official PyTorch implementation of "EcoFormer: Energy-Saving Attention with Linear Complexity"☆71Updated 2 years ago
- Code for the paper "Data Feedback Loops: Model-driven Amplification of Dataset Biases"☆15Updated 2 years ago
- Pytorch implementation for "The Surprising Positive Knowledge Transfer in Continual 3D Object Shape Reconstruction"☆33Updated 2 years ago
- PyTorch implementation of FNet: Mixing Tokens with Fourier transforms☆25Updated 3 years ago