MXHX7199 / ICCV_2021_AFPLinks
AFP is a hardware-friendly quantization framework for DNNs, which is contributed by Fangxin Liu and Wenbo Zhao.
☆12Updated 3 years ago
Alternatives and similar repositories for ICCV_2021_AFP
Users that are interested in ICCV_2021_AFP are comparing it to the libraries listed below
Sorting:
- bitfusion verilog implementation☆8Updated 3 years ago
- BSQ: Exploring Bit-Level Sparsity for Mixed-Precision Neural Network Quantization (ICLR 2021)☆40Updated 4 years ago
- ☆18Updated 2 years ago
- ☆27Updated 2 months ago
- ☆27Updated this week
- Sparse CNN Accelerator targeting Intel FPGA☆11Updated 3 years ago
- ☆18Updated 3 years ago
- A Out-of-box PyTorch Scaffold for Neural Network Quantization-Aware-Training (QAT) Research. Website: https://github.com/zhutmost/neuralz…☆26Updated 2 years ago
- LSQ+ or LSQplus☆69Updated 4 months ago
- Here are some implementations of basic hardware units in RTL language (verilog for now), which can be used for area/power evaluation and …☆11Updated last year
- ☆98Updated last year
- Open-source of MSD framework☆16Updated last year
- DeiT implementation for Q-ViT☆25Updated last month
- A bit-level sparsity-awared multiply-accumulate process element.☆16Updated 10 months ago
- A framework for fast exploration of the depth-first scheduling space for DNN accelerators☆39Updated 2 years ago
- FracBNN: Accurate and FPGA-Efficient Binary Neural Networks with Fractional Activations☆94Updated 3 years ago
- An efficient spatial accelerator enabling hybrid sparse attention mechanisms for long sequences☆27Updated last year
- ☆45Updated 3 years ago
- [TCAD'23] AccelTran: A Sparsity-Aware Accelerator for Transformers☆44Updated last year
- The official implementation of HPCA 2025 paper, Prosperity: Accelerating Spiking Neural Networks via Product Sparsity☆29Updated 4 months ago
- Post-training sparsity-aware quantization☆34Updated 2 years ago
- MINT, Multiplier-less INTeger Quantization for Energy Efficient Spiking Neural Networks, ASP-DAC 2024, Nominated for Best Paper Award☆13Updated last year
- ☆34Updated 4 years ago
- Official implementation of EMNLP'23 paper "Revisiting Block-based Quantisation: What is Important for Sub-8-bit LLM Inference?"☆22Updated last year
- LoAS: Fully Temporal-Parallel Dataflow for Dual-Sparse Spiking Neural Networks, MICRO 2024.☆11Updated 2 months ago
- A co-design architecture on sparse attention☆52Updated 3 years ago
- [ICML 2021] "Auto-NBA: Efficient and Effective Search Over the Joint Space of Networks, Bitwidths, and Accelerators" by Yonggan Fu, Yonga…☆15Updated 3 years ago
- ☆19Updated 4 years ago
- Neural Network Quantization With Fractional Bit-widths☆12Updated 4 years ago
- ☆41Updated 5 months ago