ok858ok / CP-ViT
Code for "CP-ViT: Cascade Vision Transformer Pruning via Progressive Sparsity Prediction" on CIFAR-10/100.
☆14Updated 3 years ago
Alternatives and similar repositories for CP-ViT:
Users that are interested in CP-ViT are comparing it to the libraries listed below
- Vision Transformer Pruning☆56Updated 3 years ago
- ☆26Updated this week
- [TCAD'23] AccelTran: A Sparsity-Aware Accelerator for Transformers☆40Updated last year
- ViTALiTy (HPCA'23) Code Repository☆22Updated 2 years ago
- ☆18Updated 3 years ago
- [HPCA'21] SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning☆84Updated 7 months ago
- ☆43Updated 3 years ago
- AFP is a hardware-friendly quantization framework for DNNs, which is contributed by Fangxin Liu and Wenbo Zhao.☆12Updated 3 years ago
- An efficient spatial accelerator enabling hybrid sparse attention mechanisms for long sequences☆26Updated last year
- ☆26Updated 3 weeks ago
- A co-design architecture on sparse attention☆52Updated 3 years ago
- BSQ: Exploring Bit-Level Sparsity for Mixed-Precision Neural Network Quantization (ICLR 2021)☆40Updated 4 years ago
- Open-source of MSD framework☆16Updated last year
- ☆34Updated 4 years ago
- DeiT implementation for Q-ViT☆24Updated this week
- ☆15Updated 2 years ago
- ☆18Updated 2 years ago
- [HPCA 2023] ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design☆105Updated last year
- A FPGA-based neural network inference accelerator, which won the third place in DAC-SDC☆28Updated 2 years ago
- A Out-of-box PyTorch Scaffold for Neural Network Quantization-Aware-Training (QAT) Research. Website: https://github.com/zhutmost/neuralz…☆26Updated 2 years ago
- BitSplit Post-trining Quantization☆49Updated 3 years ago
- Vision Transformer Accelerator implemented in Vivado HLS for Xilinx FPGAs.☆11Updated 3 months ago
- Here are some implementations of basic hardware units in RTL language (verilog for now), which can be used for area/power evaluation and …☆11Updated last year
- Model LLM inference on single-core dataflow accelerators☆10Updated 2 months ago
- ☆12Updated last year
- ☆95Updated last year
- Training with Block Minifloat number representation☆14Updated 3 years ago
- ☆43Updated 2 years ago
- A bit-level sparsity-awared multiply-accumulate process element.☆14Updated 9 months ago
- The final project repository for 2022 Spring COMS6998-009 Deep Learning System Performance in Columbia University.☆7Updated 2 years ago