cmsflash / efficient-attention
An implementation of the efficient attention module.
☆301Updated 4 years ago
Alternatives and similar repositories for efficient-attention:
Users that are interested in efficient-attention are comparing it to the libraries listed below
- Implementation of Transformer in Transformer, pixel level attention paired with patch level attention for image classification, in Pytorc…☆305Updated 3 years ago
- ☆191Updated 2 years ago
- [ICLR'22 Oral] Implementation of "CycleMLP: A MLP-like Architecture for Dense Prediction"☆283Updated 2 years ago
- [NeurIPS 2021] [T-PAMI] Global Filter Networks for Image Classification☆465Updated last year
- [NeurIPS 2021 Spotlight] Official code for "Focal Self-attention for Local-Global Interactions in Vision Transformers"☆549Updated 2 years ago
- ☆245Updated 2 years ago
- Implementation of ConvMixer for "Patches Are All You Need? 🤷"☆1,064Updated 2 years ago
- Implementation of Axial attention - attending to multi-dimensional data efficiently☆371Updated 3 years ago
- MLP-Like Vision Permutator for Visual Recognition (PyTorch)☆191Updated 2 years ago
- Implementation of Linformer for Pytorch☆266Updated last year
- Unofficial implementation of MLP-Mixer: An all-MLP Architecture for Vision☆217Updated 3 years ago
- Official code for paper "On the Connection between Local Attention and Dynamic Depth-wise Convolution" ICLR 2022 Spotlight☆184Updated 2 years ago
- Code repository of the paper "Modelling Long Range Dependencies in ND: From Task-Specific to a General Purpose CNN" https://arxiv.org/abs…☆183Updated last year
- Unofficial implementation of Google's FNet: Mixing Tokens with Fourier Transforms☆258Updated 3 years ago
- Official PyTorch Implementation of Long-Short Transformer (NeurIPS 2021).☆226Updated 2 years ago
- Transformer based on a variant of attention that is linear complexity in respect to sequence length☆738Updated 9 months ago
- MetaFormer Baselines for Vision (TPAMI 2024)☆443Updated 8 months ago
- Implementation of the 😇 Attention layer from the paper, Scaling Local Self-Attention For Parameter Efficient Visual Backbones☆198Updated 3 years ago
- [ICLR 2021 top 3%] Is Attention Better Than Matrix Decomposition?☆330Updated 2 years ago
- Implementation of Pixel-level Contrastive Learning, proposed in the paper "Propagate Yourself", in Pytorch☆257Updated 4 years ago
- iFormer: Inception Transformer☆246Updated 2 years ago
- Neighborhood Attention Transformer, arxiv 2022 / CVPR 2023. Dilated Neighborhood Attention Transformer, arxiv 2022☆1,088Updated 9 months ago
- ConvMAE: Masked Convolution Meets Masked Autoencoders☆491Updated last year
- PyTorch Implementation of CvT: Introducing Convolutions to Vision Transformers☆227Updated 3 years ago
- ☆197Updated 6 months ago
- Fully featured implementation of Routing Transformer☆288Updated 3 years ago
- A better PyTorch implementation of image local attention which reduces the GPU memory by an order of magnitude.☆137Updated 3 years ago
- This is an official implementation of CvT: Introducing Convolutions to Vision Transformers.☆566Updated last year
- [NeurIPS 2022] HorNet: Efficient High-Order Spatial Interactions with Recursive Gated Convolutions☆327Updated last year
- Self-supervised vIsion Transformer (SiT)☆325Updated 2 years ago