[NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang, Zhangyang Wang
☆89Dec 1, 2023Updated 2 years ago
Alternatives and similar repositories for SViTE
Users that are interested in SViTE are comparing it to the libraries listed below
Sorting:
- [ICLR 2022] "Unified Vision Transformer Compression" by Shixing Yu*, Tianlong Chen*, Jiayi Shen, Huan Yuan, Jianchao Tan, Sen Yang, Ji Li…☆55Dec 1, 2023Updated 2 years ago
- ☆21Apr 24, 2022Updated 3 years ago
- Revisiting Parameter Sharing for Automatic Neural Channel Number Search, NeurIPS 2020☆22Nov 15, 2020Updated 5 years ago
- Code for "CP-ViT: Cascade Vision Transformer Pruning via Progressive Sparsity Prediction" on CIFAR-10/100.☆14Dec 10, 2021Updated 4 years ago
- ☆53Aug 28, 2024Updated last year
- [NeurIPS 2021] [T-PAMI] DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification☆651Jul 11, 2023Updated 2 years ago
- Code for ViTAS_Vision Transformer Architecture Search☆51Jul 22, 2021Updated 4 years ago
- Official implement of Evo-ViT: Slow-Fast Token Evolution for Dynamic Vision Transformer☆74Jul 13, 2022Updated 3 years ago
- (CVPR 2021, Oral) Dynamic Slimmable Network☆231Dec 31, 2021Updated 4 years ago
- (ICCV 2021) BossNAS: Exploring Hybrid CNN-transformers with Block-wisely Self-supervised Neural Architecture Search☆142Dec 6, 2021Updated 4 years ago
- This repository contains the code for the paper in Findings of EMNLP 2021: "EfficientBERT: Progressively Searching Multilayer Perceptron …☆33Jun 14, 2023Updated 2 years ago
- (ACL-IJCNLP 2021) Convolutions and Self-Attention: Re-interpreting Relative Positions in Pre-trained Language Models.☆21Jul 13, 2022Updated 3 years ago
- Neural Network Quantization With Fractional Bit-widths☆11Feb 19, 2021Updated 5 years ago
- A pytorch implementation of the ICCV2021 workshop paper SimDis: Simple Distillation Baselines for Improving Small Self-supervised Models☆14Jul 15, 2021Updated 4 years ago
- (ICCV 2021 Oral) CoaT: Co-Scale Conv-Attentional Image Transformers☆235Feb 3, 2022Updated 4 years ago
- BitSplit Post-trining Quantization☆50Dec 20, 2021Updated 4 years ago
- HR-NAS: Searching Efficient High-Resolution Neural Architectures with Lightweight Transformers (CVPR21 Oral)☆143Jun 14, 2021Updated 4 years ago
- [NeurIPS 2021] “Stronger NAS with Weaker Predictors“, Junru Wu, Xiyang Dai, Dongdong Chen, Yinpeng Chen, Mengchen Liu, Ye Yu, Zhangyang W…☆27Sep 23, 2022Updated 3 years ago
- Bag of MLP☆20May 31, 2021Updated 4 years ago
- Pruning Neural Networks with Taylor criterion in Pytorch☆321Nov 3, 2019Updated 6 years ago
- MLP-Like Vision Permutator for Visual Recognition (PyTorch)☆192Mar 31, 2022Updated 3 years ago
- [ICML 2021] "Auto-NBA: Efficient and Effective Search Over the Joint Space of Networks, Bitwidths, and Accelerators" by Yonggan Fu, Yonga…☆16Jan 3, 2022Updated 4 years ago
- BESA is a differentiable weight pruning technique for large language models.☆17Mar 4, 2024Updated 2 years ago
- Successfully training approximations to full-rank matrices for efficiency in deep learning.☆17Jan 5, 2021Updated 5 years ago
- [TPAMI 2024] This is the official repository for our paper: ''Pruning Self-attentions into Convolutional Layers in Single Path''.☆115Dec 30, 2023Updated 2 years ago
- Official code for our CVPR'22 paper “Vision Transformer Slimming: Multi-Dimension Searching in Continuous Optimization Space”☆251Aug 24, 2025Updated 6 months ago
- [CVPR 2021] "The Lottery Tickets Hypothesis for Supervised and Self-supervised Pre-training in Computer Vision Models" Tianlong Chen, Jon…☆68Dec 17, 2022Updated 3 years ago
- ResRep: Lossless CNN Pruning via Decoupling Remembering and Forgetting (ICCV 2021)☆300Dec 1, 2022Updated 3 years ago
- The official implementation of the NeurIPS 2022 paper Q-ViT.☆105May 22, 2023Updated 2 years ago
- [NeurIPS 2019] E2-Train: Training State-of-the-art CNNs with Over 80% Less Energy☆21Nov 18, 2019Updated 6 years ago
- [KDD'22] Learned Token Pruning for Transformers☆101Feb 27, 2023Updated 3 years ago
- [ICLR 2022] "Learning Pruning-Friendly Networks via Frank-Wolfe: One-Shot, Any-Sparsity, and No Retraining" by Lu Miao*, Xiaolong Luo*, T…☆33Jan 20, 2022Updated 4 years ago
- Implementation of a Quantized Transformer Model☆19Mar 20, 2019Updated 6 years ago
- [CVPRW 21] "BNN - BN = ? Training Binary Neural Networks without Batch Normalization", Tianlong Chen, Zhenyu Zhang, Xu Ouyang, Zechun Liu…☆57Dec 30, 2021Updated 4 years ago
- ☆98Apr 27, 2022Updated 3 years ago
- [ICLR 2020]: 'AtomNAS: Fine-Grained End-to-End Neural Architecture Search'☆220Jun 8, 2020Updated 5 years ago
- Code for "Picking Winning Tickets Before Training by Preserving Gradient Flow" https://openreview.net/pdf?id=SkgsACVKPH☆105Feb 18, 2020Updated 6 years ago
- [NeurIPS‘2021] "MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge", Geng Yuan, Xiaolong Ma, Yanzhi Wang et al…☆17Mar 16, 2022Updated 3 years ago
- Vision Transformer Pruning☆57Dec 9, 2021Updated 4 years ago