VITA-Group / AsViT
[ICLR 2022] "As-ViT: Auto-scaling Vision Transformers without Training" by Wuyang Chen, Wei Huang, Xianzhi Du, Xiaodan Song, Zhangyang Wang, Denny Zhou
☆76Updated 3 years ago
Alternatives and similar repositories for AsViT:
Users that are interested in AsViT are comparing it to the libraries listed below
- code for NASViT☆67Updated 2 years ago
- Code for ViTAS_Vision Transformer Architecture Search☆51Updated 3 years ago
- [NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang…☆89Updated last year
- (CVPR 2022) Automated Progressive Learning for Efficient Training of Vision Transformers☆25Updated last month
- ☆57Updated 3 years ago
- This is a offical PyTorch/GPU implementation of SupMAE.☆77Updated 2 years ago
- [ICLR 2022] "Unified Vision Transformer Compression" by Shixing Yu*, Tianlong Chen*, Jiayi Shen, Huan Yuan, Jianchao Tan, Sen Yang, Ji Li…☆52Updated last year
- This is the official PyTorch implementation for "Mesa: A Memory-saving Training Framework for Transformers".☆120Updated 3 years ago
- code release of research paper "Exploring Long-Sequence Masked Autoencoders"☆100Updated 2 years ago
- [NeurIPS 2022 Spotlight] This is the official PyTorch implementation of "EcoFormer: Energy-Saving Attention with Linear Complexity"☆71Updated 2 years ago
- Cyclic Differentiable Architecture Search☆36Updated 3 years ago
- Bag of Instances Aggregation Boosts Self-supervised Distillation (ICLR 2022)☆33Updated 2 years ago
- Code and models for the paper Glance-and-Gaze Vision Transformer☆28Updated 3 years ago
- Beyond Masking: Demystifying Token-Based Pre-Training for Vision Transformers☆26Updated 3 years ago
- [ICML 2022] "DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks", by Yonggan …☆71Updated 2 years ago
- Benchmarking Attention Mechanism in Vision Transformers.☆17Updated 2 years ago
- ☆73Updated 2 years ago
- [NeurIPS'22] What Makes a "Good" Data Augmentation in Knowledge Distillation -- A Statistical Perspective☆36Updated 2 years ago
- [CVPR 2021] "The Lottery Tickets Hypothesis for Supervised and Self-supervised Pre-training in Computer Vision Models" Tianlong Chen, Jon…☆69Updated 2 years ago
- [NeurIPS 2021] ORL: Unsupervised Object-Level Representation Learning from Scene Images☆58Updated 3 years ago
- Official Pytorch implementation of Super Vision Transformer (IJCV)☆43Updated last year
- code for "AttentiveNAS Improving Neural Architecture Search via Attentive Sampling"☆104Updated 3 years ago
- Learning recognition/segmentation models without end-to-end training. 40%-60% less GPU memory footprint. Same training time. Better perfo…☆90Updated 2 years ago
- ☆50Updated last year
- Official Pytorch implementation for Distilling Image Classifiers in Object detection (NeurIPS2021)☆31Updated 3 years ago
- ☆108Updated 3 years ago
- Unified Architecture Search with Convolution, Transformer, and MLP (ECCV 2022)☆53Updated 2 years ago
- [ICLR'23] Trainability Preserving Neural Pruning (PyTorch)☆33Updated last year
- UniMoCo: Unsupervised, Semi-Supervised and Full-Supervised Visual Representation Learning☆55Updated 3 years ago
- ☆16Updated last year