mit-han-lab / sparsevitLinks
[CVPR'23] SparseViT: Revisiting Activation Sparsity for Efficient High-Resolution Vision Transformer
☆70Updated last year
Alternatives and similar repositories for sparsevit
Users that are interested in sparsevit are comparing it to the libraries listed below
Sorting:
- [ICCV 23]An approach to enhance the efficiency of Vision Transformer (ViT) by concurrently employing token pruning and token merging tech…☆96Updated last year
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆30Updated last year
- [BMVC 2024] PlainMamba: Improving Non-hierarchical Mamba in Visual Recognition☆78Updated 2 months ago
- Official PyTorch implementation of Which Tokens to Use? Investigating Token Reduction in Vision Transformers presented at ICCV 2023 NIVT …☆34Updated last year
- This is the official pytorch implementation for the paper: Towards Accurate Post-training Quantization for Diffusion Models.(CVPR24 Poste…☆37Updated last year
- Official repository of InLine attention (NeurIPS 2024)☆48Updated 6 months ago
- ☆45Updated last year
- [ECCV 2024] Isomorphic Pruning for Vision Models☆69Updated 11 months ago
- [ICLR 2025] Mixture Compressor for Mixture-of-Experts LLMs Gains More☆46Updated 4 months ago
- Official implementation of "SViT: Revisiting Token Pruning for Object Detection and Instance Segmentation"☆32Updated last year
- ALGM applied to Segmenter☆26Updated last year
- [NeurIPS2024 Spotlight] The official implementation of MambaTree: Tree Topology is All You Need in State Space Model☆93Updated last year
- [ICLR2025] This repository is the official implementation of our Autoregressive Pretraining with Mamba in Vision☆80Updated 3 weeks ago
- [ECCV 2024 Workshop Best Paper Award] Famba-V: Fast Vision Mamba with Cross-Layer Token Fusion☆34Updated 8 months ago
- ☆12Updated last year
- [ICLR 2024] This is the official PyTorch implementation of "QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Mod…☆38Updated last year
- An official code release of the paper RGB no more: Minimally Decoded JPEG Vision Transformers☆55Updated last year
- Adapting LLaMA Decoder to Vision Transformer☆28Updated last year
- [NeurIPS 2022] “M³ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task Learning with Model-Accelerator Co-design”, Hanxue …☆121Updated 2 years ago
- Code release for Deep Incubation (https://arxiv.org/abs/2212.04129)☆90Updated 2 years ago
- [NeurIPS'24]Efficient and accurate memory saving method towards W4A4 large multi-modal models.☆74Updated 5 months ago
- [ICLR 2024 Spotlight] This is the official PyTorch implementation of "EfficientDM: Efficient Quantization-Aware Fine-Tuning of Low-Bit Di…☆60Updated last year
- Open source implementation of "Vision Transformers Need Registers"☆182Updated 2 months ago
- torch_quantizer is a out-of-box quantization tool for PyTorch models on CUDA backend, specially optimized for Diffusion Models.☆22Updated last year
- [NeurIPS 2023] ShiftAddViT: Mixture of Multiplication Primitives Towards Efficient Vision Transformer☆31Updated last year
- [CVPR-22] This is the official implementation of the paper "Adavit: Adaptive vision transformers for efficient image recognition".☆54Updated 2 years ago
- The official implementation of "2024NeurIPS Dynamic Tuning Towards Parameter and Inference Efficiency for ViT Adaptation"☆46Updated 5 months ago
- A library for calculating the FLOPs in the forward() process based on torch.fx☆115Updated 2 months ago
- The codebase for paper "PPT: Token Pruning and Pooling for Efficient Vision Transformer"☆23Updated 7 months ago
- PyTorch implementation of PTQ4DiT https://arxiv.org/abs/2405.16005☆30Updated 7 months ago