tobna / WhatTransformerToFavor
Github repository for the paper Which Transformer to Favor: A Comparative Analysis of Efficiency in Vision Transformers.
☆28Updated last month
Alternatives and similar repositories for WhatTransformerToFavor:
Users that are interested in WhatTransformerToFavor are comparing it to the libraries listed below
- ☆24Updated last year
- HSViT: Horizontally Scalable Vision Transformer☆13Updated 6 months ago
- [ICCV W] Contextual Convolutional Neural Networks (https://arxiv.org/pdf/2108.07387.pdf)☆13Updated 3 years ago
- Official implementation of ParCNetV2☆10Updated last year
- Unofficial pytorch implementation of Piecewise Linear Unit dynamic activation function☆17Updated 2 years ago
- The official project website of "NORM: Knowledge Distillation via N-to-One Representation Matching" (The paper of NORM is published in IC…☆20Updated last year
- ☆14Updated 2 years ago
- The codebase for paper "PPT: Token Pruning and Pooling for Efficient Vision Transformer"☆23Updated 5 months ago
- This is an official implementation of our CVPR 2020 paper "Non-Local Neural Networks With Grouped Bilinear Attentional Transforms".☆12Updated 4 years ago
- Pytorch implementation of our paper accepted by ECCV2022 -- Knowledge Condensation Distillation https://arxiv.org/abs/2207.05409☆30Updated 2 years ago
- ☆45Updated last year
- ☆22Updated 3 years ago
- Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation. NeurIPS 2022.☆32Updated 2 years ago
- [ECCV 2022] Implementation of the paper "Locality Guidance for Improving Vision Transformers on Tiny Datasets"☆79Updated 2 years ago
- [ECCV 2022] AMixer: Adaptive Weight Mixing for Self-attention Free Vision Transformers☆28Updated 2 years ago
- ResMLP: Feedforward networks for image classification with data-efficient training☆42Updated 3 years ago
- ☆12Updated 2 years ago
- Knowledge Distillation Toolbox for Semantic Segmentation☆17Updated 2 years ago
- [ICML 2022] This work investigates the compatibility between label smoothing (LS) and knowledge distillation (KD). We suggest to use an L…☆11Updated 2 years ago
- [ICCV 2023] Source code of "Fcaformer: Forward Cross Attention in Hybrid Vision Transformer"☆22Updated last year
- Scattering Vision Transformer☆50Updated last year
- Official code release of our paper "FViT: A Focal Vision Transformer with Gabor Filter"☆16Updated last year
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆30Updated last year
- A Simple Adaptive Unfolding Network for Hyperspectral Image Reconstruction☆33Updated 2 years ago
- Pytorch implementation (TPAMI 2023) - Training Compact CNNs for Image Classification using Dynamic-coded Filter Fusion☆19Updated 2 years ago
- ☆43Updated 2 years ago
- Training ImageNet / CIFAR models with sota strategies and fancy techniques such as ViT, KD, Rep, etc.☆82Updated last year
- Code for Learned Thresholds Token Merging and Pruning for Vision Transformers (LTMP). A technique to reduce the size of Vision Transforme…☆16Updated 5 months ago
- This is the official code for paper: Token Summarisation for Efficient Vision Transformers via Graph-based Token Propagation☆27Updated last year
- Official PyTorch implementation of ResFormer: Scaling ViTs with Multi-Resolution Training, CVPR2023☆27Updated last year