pprp / Vision-Mamba-CIFAR10
☆24Updated last year
Alternatives and similar repositories for Vision-Mamba-CIFAR10:
Users that are interested in Vision-Mamba-CIFAR10 are comparing it to the libraries listed below
- ☆26Updated last year
- To appear in the 11th International Conference on Learning Representations (ICLR 2023).☆17Updated 2 years ago
- Source code of our TNNLS paper "Boosting Convolutional Neural Networks with Middle Spectrum Grouped Convolution"☆12Updated 2 years ago
- The official project website of "NORM: Knowledge Distillation via N-to-One Representation Matching" (The paper of NORM is published in IC…☆20Updated last year
- Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation. NeurIPS 2022.☆32Updated 2 years ago
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆30Updated last year
- Training ImageNet / CIFAR models with sota strategies and fancy techniques such as ViT, KD, Rep, etc.☆82Updated last year
- ☆8Updated last year
- Unofficial pytorch implementation of Piecewise Linear Unit dynamic activation function☆17Updated 2 years ago
- ☆26Updated 2 years ago
- Code for 'Multi-level Logit Distillation' (CVPR2023)☆63Updated 7 months ago
- ☆45Updated last year
- Implementation of PGONAS for CVPR22W and RD-NAS for ICASSP23☆22Updated 2 years ago
- The codebase for paper "PPT: Token Pruning and Pooling for Efficient Vision Transformer"☆23Updated 5 months ago
- This is the official code for paper: Token Summarisation for Efficient Vision Transformers via Graph-based Token Propagation☆27Updated last year
- ☆12Updated last year
- Auto-Prox-AAAI24☆12Updated last year
- Github repository for the paper Which Transformer to Favor: A Comparative Analysis of Efficiency in Vision Transformers.☆28Updated last month
- [TPAMI-2023] Official implementations of L-MCL: Online Knowledge Distillation via Mutual Contrastive Learning for Visual Recognition☆23Updated last year
- Implementation of HAT https://arxiv.org/pdf/2204.00993☆50Updated last year
- Code for Learned Thresholds Token Merging and Pruning for Vision Transformers (LTMP). A technique to reduce the size of Vision Transforme…☆16Updated 5 months ago
- Official code for Scale Decoupled Distillation☆41Updated last year
- ☆47Updated 8 months ago
- Pytorch implementation of our paper accepted by ECCV2022 -- Knowledge Condensation Distillation https://arxiv.org/abs/2207.05409☆30Updated 2 years ago
- [ICCV-2023] EMQ: Evolving Training-free Proxies for Automated Mixed Precision Quantization☆25Updated last year
- ☆11Updated last year
- [NeurIPS2023]Lightweight Vision Transformer with Bidirectional Interaction☆23Updated last year
- TF-FD☆20Updated 2 years ago
- HSViT: Horizontally Scalable Vision Transformer☆13Updated 6 months ago
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated 2 years ago