tobna / WhatTransformerToFavor
Github repository for the paper Which Transformer to Favor: A Comparative Analysis of Efficiency in Vision Transformers.
☆25Updated 4 months ago
Alternatives and similar repositories for WhatTransformerToFavor:
Users that are interested in WhatTransformerToFavor are comparing it to the libraries listed below
- Official implementation of ParCNetV2☆10Updated 11 months ago
- Official PyTorch implementation of ResFormer: Scaling ViTs with Multi-Resolution Training, CVPR2023☆25Updated last year
- [CVPR2023] Practical Network Acceleration with Tiny Sets☆12Updated last year
- ☆34Updated last year
- ☆32Updated last year
- Official code for BA-SAM:Scalable Bias-Mode Attention Mask for Segment Anything Model☆16Updated 7 months ago
- EATFormer: Improving Vision Transformer Inspired by Evolutionary Algorithm☆32Updated 2 years ago
- [ECCV 2022] Implementation of the paper "Locality Guidance for Improving Vision Transformers on Tiny Datasets"☆78Updated 2 years ago
- [CVPR 2024] VkD : Improving Knowledge Distillation using Orthogonal Projections☆50Updated 4 months ago
- ResMLP: Feedforward networks for image classification with data-efficient training☆42Updated 3 years ago
- ☆21Updated 3 years ago
- [ECCV 2022] AMixer: Adaptive Weight Mixing for Self-attention Free Vision Transformers☆28Updated 2 years ago
- Unified Architecture Search with Convolution, Transformer, and MLP (ECCV 2022)☆53Updated 2 years ago
- Source code of our TNNLS paper "Boosting Convolutional Neural Networks with Middle Spectrum Grouped Convolution"☆10Updated last year
- The official implementation for ALOFT (CVPR 2023).☆53Updated last year
- ☆18Updated 4 months ago
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆28Updated 11 months ago
- Pytorch implementation of our paper accepted by ECCV2022 -- Knowledge Condensation Distillation https://arxiv.org/abs/2207.05409☆30Updated 2 years ago
- [BMVC 2024] PlainMamba: Improving Non-hierarchical Mamba in Visual Recognition☆73Updated 6 months ago
- ☆31Updated 5 months ago
- The official implementation of "Adapter is All You Need for Tuning Visual Tasks".☆77Updated 5 months ago
- A Close Look at Spatial Modeling: From Attention to Convolution☆91Updated 2 years ago
- ☆52Updated last year
- The official project website of "NORM: Knowledge Distillation via N-to-One Representation Matching" (The paper of NORM is published in IC…☆19Updated last year
- Official implement of Evo-ViT: Slow-Fast Token Evolution for Dynamic Vision Transformer☆71Updated 2 years ago
- ☆58Updated 2 years ago
- This is code of paper "ScalableViT: Rethinking the Context-oriented Generalization of Vision Transformer"☆26Updated last year
- Visual Dependency Transformers: Dependency Tree Emerges from Reversed Attention (CVPR 2023)☆32Updated last year
- Knowledge Distillation Toolbox for Semantic Segmentation☆17Updated 2 years ago