VITA-Group / AsViT
[ICLR 2022] "As-ViT: Auto-scaling Vision Transformers without Training" by Wuyang Chen, Wei Huang, Xianzhi Du, Xiaodan Song, Zhangyang Wang, Denny Zhou
☆76Updated 3 years ago
Alternatives and similar repositories for AsViT:
Users that are interested in AsViT are comparing it to the libraries listed below
- code for NASViT☆67Updated 3 years ago
- (CVPR 2022) Automated Progressive Learning for Efficient Training of Vision Transformers☆25Updated 2 months ago
- ☆57Updated 3 years ago
- Code for ViTAS_Vision Transformer Architecture Search☆51Updated 3 years ago
- [NeurIPS 2022 Spotlight] This is the official PyTorch implementation of "EcoFormer: Energy-Saving Attention with Linear Complexity"☆71Updated 2 years ago
- [NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang…☆89Updated last year
- This is a offical PyTorch/GPU implementation of SupMAE.☆78Updated 2 years ago
- [ICLR 2022] "Unified Vision Transformer Compression" by Shixing Yu*, Tianlong Chen*, Jiayi Shen, Huan Yuan, Jianchao Tan, Sen Yang, Ji Li…☆52Updated last year
- Cyclic Differentiable Architecture Search☆36Updated 3 years ago
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated 2 years ago
- [ICLR'23] Trainability Preserving Neural Pruning (PyTorch)☆33Updated last year
- PyTorch implementation of SSQL (Accepted to ECCV2022 oral presentation)☆74Updated 2 years ago
- This is the official PyTorch implementation for "Mesa: A Memory-saving Training Framework for Transformers".☆120Updated 3 years ago
- [ICML 2022] "DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks", by Yonggan …☆71Updated 2 years ago
- [ICLR 2022] The Unreasonable Effectiveness of Random Pruning: Return of the Most Naive Baseline for Sparse Training by Shiwei Liu, Tianlo…☆73Updated 2 years ago
- Code and models for the paper Glance-and-Gaze Vision Transformer☆28Updated 3 years ago
- [CVPR 2021] "The Lottery Tickets Hypothesis for Supervised and Self-supervised Pre-training in Computer Vision Models" Tianlong Chen, Jon…☆69Updated 2 years ago
- Bag of Instances Aggregation Boosts Self-supervised Distillation (ICLR 2022)☆33Updated 3 years ago
- [ICME 2022] code for the paper, SimVit: Exploring a simple vision transformer with sliding windows.☆68Updated 2 years ago
- ☆73Updated 2 years ago
- Official Pytorch implementation for Distilling Image Classifiers in Object detection (NeurIPS2021)☆31Updated 3 years ago
- code for "AttentiveNAS Improving Neural Architecture Search via Attentive Sampling"☆104Updated 3 years ago
- Learning recognition/segmentation models without end-to-end training. 40%-60% less GPU memory footprint. Same training time. Better perfo…☆90Updated 2 years ago
- [NeurIPS 2021] ORL: Unsupervised Object-Level Representation Learning from Scene Images☆58Updated 3 years ago
- [ICML 2021 Oral] "CATE: Computation-aware Neural Architecture Encoding with Transformers" by Shen Yan, Kaiqiang Song, Fei Liu, Mi Zhang☆19Updated 3 years ago
- A compilation of network architectures for vision and others without usage of self-attention mechanism☆80Updated 2 years ago
- ☆16Updated 2 years ago
- AlphaNet Improved Training of Supernet with Alpha-Divergence☆98Updated 3 years ago
- Metrics for "Beyond neural scaling laws: beating power law scaling via data pruning " (NeurIPS 2022 Outstanding Paper Award)☆56Updated 2 years ago
- Benchmarking Attention Mechanism in Vision Transformers.☆17Updated 2 years ago