lucidrains / transformer-in-transformer
Implementation of Transformer in Transformer, pixel level attention paired with patch level attention for image classification, in Pytorch
β300Updated 2 years ago
Related projects β
Alternatives and complementary repositories for transformer-in-transformer
- Implementation of the π Attention layer from the paper, Scaling Local Self-Attention For Parameter Efficient Visual Backbonesβ199Updated 3 years ago
- [ICLR'22 Oral] Implementation of "CycleMLP: A MLP-like Architecture for Dense Prediction"β280Updated 2 years ago
- Implementation of Pixel-level Contrastive Learning, proposed in the paper "Propagate Yourself", in Pytorchβ252Updated 3 years ago
- [NeurIPS 2021 Spotlight] Official code for "Focal Self-attention for Local-Global Interactions in Vision Transformers"β544Updated 2 years ago
- Self-supervised vIsion Transformer (SiT)β324Updated last year
- PyTorch Implementation of CvT: Introducing Convolutions to Vision Transformersβ224Updated 3 years ago
- EsViT: Efficient self-supervised Vision Transformersβ408Updated last year
- β189Updated last year
- Dense Contrastive Learning (DenseCL) for self-supervised representation learning, CVPR 2021 Oral.β546Updated 10 months ago
- This is an official implementation for "Self-Supervised Learning with Swin Transformers".β624Updated 3 years ago
- β239Updated 2 years ago
- Propagate Yourself: Exploring Pixel-Level Consistency for Unsupervised Visual Representation Learning, CVPR 2021β332Updated 3 years ago
- Unofficial implementation of MLP-Mixer: An all-MLP Architecture for Visionβ209Updated 3 years ago
- This is a PyTorch re-implementation of Axial-DeepLab (ECCV 2020 Spotlight)β449Updated 3 years ago
- Pytorch implementation of "All Tokens Matter: Token Labeling for Training Better Vision Transformers"β426Updated last year
- Code for the Convolutional Vision Transformer (ConViT)β463Updated 3 years ago
- A Pytorch Implementation of the following paper "Visual Transformers: Token-based Image Representation and Processing for Computer Visionβ¦β181Updated 3 years ago
- Two simple and effective designs of vision transformer, which is on par with the Swin transformerβ580Updated last year
- (ICCV 2021 Oral) CoaT: Co-Scale Conv-Attentional Image Transformersβ228Updated 2 years ago
- MLP-Like Vision Permutator for Visual Recognition (PyTorch)β190Updated 2 years ago
- An implementation of the efficient attention module.β283Updated 3 years ago
- This is an official implementation for "ResT: An Efficient Transformer for Visual Recognition".β280Updated 2 years ago
- Bottleneck Transformers for Visual Recognitionβ274Updated 3 years ago
- An Image is Worth 16x16 Words: Transformers for Image Recognition at Scaleβ287Updated 3 years ago
- ConvMAE: Masked Convolution Meets Masked Autoencodersβ483Updated last year
- Pytorch version of Vision Transformer (ViT) with pretrained models. This is part of CASL (https://casl-project.github.io/) and ASYML projβ¦β342Updated 3 years ago
- Implementation of Axial attention - attending to multi-dimensional data efficientlyβ351Updated 3 years ago
- β243Updated 3 years ago
- Implementing Lambda Networks using Pytorchβ138Updated 3 years ago
- Implementation of CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classificationβ188Updated 3 years ago