tobna / WhatTransformerToFavorLinks
Github repository for the paper Which Transformer to Favor: A Comparative Analysis of Efficiency in Vision Transformers.
☆28Updated 2 months ago
Alternatives and similar repositories for WhatTransformerToFavor
Users that are interested in WhatTransformerToFavor are comparing it to the libraries listed below
Sorting:
- The official project website of "NORM: Knowledge Distillation via N-to-One Representation Matching" (The paper of NORM is published in IC…☆20Updated last year
- ResMLP: Feedforward networks for image classification with data-efficient training☆42Updated 4 years ago
- Official implementation of the paper "Masked Autoencoders are Efficient Class Incremental Learners"☆42Updated last year
- This is an official implementation of our NeurIPS 2022 paper "Bridging the Gap Between Vision Transformers and Convolutional Neural Netwo…☆56Updated 2 years ago
- [CVPR 2024] VkD : Improving Knowledge Distillation using Orthogonal Projections☆53Updated 7 months ago
- HSViT: Horizontally Scalable Vision Transformer☆13Updated 7 months ago
- Source code of our TNNLS paper "Boosting Convolutional Neural Networks with Middle Spectrum Grouped Convolution"☆12Updated 2 years ago
- The offical implementation of [NeurIPS2024] Wasserstein Distance Rivals Kullback-Leibler Divergence for Knowledge Distillation https://ar…☆39Updated 5 months ago
- Official implementation of ParCNetV2☆10Updated last year
- Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation. NeurIPS 2022.☆32Updated 2 years ago
- ☆23Updated last year
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆30Updated last year
- Information Bottleneck Approach to Spatial Attention Learning, IJCAI2021☆15Updated 4 years ago
- [ICML 2022] This work investigates the compatibility between label smoothing (LS) and knowledge distillation (KD). We suggest to use an L…☆11Updated 2 years ago
- Training ImageNet / CIFAR models with sota strategies and fancy techniques such as ViT, KD, Rep, etc.☆82Updated last year
- Pytorch implementation of our paper accepted by ECCV2022 -- Knowledge Condensation Distillation https://arxiv.org/abs/2207.05409☆30Updated 2 years ago
- Official implement of Evo-ViT: Slow-Fast Token Evolution for Dynamic Vision Transformer☆72Updated 2 years ago
- ☆45Updated last year
- ☆22Updated 3 years ago
- Pytorch implementation (TPAMI 2023) - Training Compact CNNs for Image Classification using Dynamic-coded Filter Fusion☆19Updated 2 years ago
- official pytorch implementation of Tobias: A random CNN sees objects and its applications in self-supervised learning.☆28Updated 3 years ago
- ☆14Updated 2 years ago
- ☆11Updated last year
- ☆50Updated 9 months ago
- The codebase for paper "PPT: Token Pruning and Pooling for Efficient Vision Transformer"☆23Updated 6 months ago
- [ECCV 2022] AMixer: Adaptive Weight Mixing for Self-attention Free Vision Transformers☆28Updated 2 years ago
- ☆57Updated 2 years ago
- ☆19Updated last year
- [ICCV W] Contextual Convolutional Neural Networks (https://arxiv.org/pdf/2108.07387.pdf)☆13Updated 3 years ago
- [ICLR 2022] "Anti-Oversmoothing in Deep Vision Transformers via the Fourier Domain Analysis: From Theory to Practice" by Peihao Wang, Wen…☆79Updated last year