e-dupuis / awesome-approximate-dnn
Curated content for DNN approximation, acceleration ... with a focus on hardware accelerator and deployment
☆23Updated 4 months ago
Related projects: ⓘ
- [ICML 2021] "Auto-NBA: Efficient and Effective Search Over the Joint Space of Networks, Bitwidths, and Accelerators" by Yonggan Fu, Yonga…☆14Updated 2 years ago
- [ICASSP'20] DNN-Chip Predictor: An Analytical Performance Predictor for DNN Accelerators with Various Dataflows and Hardware Architecture…☆22Updated last year
- [FPGA'21] CoDeNet is an efficient object detection model on PyTorch, with SOTA performance on VOC and COCO based on CenterNet and Co-Desi…☆25Updated last year
- A Out-of-box PyTorch Scaffold for Neural Network Quantization-Aware-Training (QAT) Research. Website: https://github.com/zhutmost/neuralz…☆26Updated last year
- [TCAD 2021] Block Convolution: Towards Memory-Efficient Inference of Large-Scale CNNs on FPGA☆15Updated 2 years ago
- ☆17Updated last year
- A DAG processor and compiler for a tree-based spatial datapath.☆12Updated 2 years ago
- ☆23Updated 2 years ago
- A FPGA-based neural network inference accelerator, which won the third place in DAC-SDC☆28Updated 2 years ago
- HW/SW co-design of sentence-level energy optimizations for latency-aware multi-task NLP inference☆46Updated 5 months ago
- ☆19Updated 3 years ago
- ☆23Updated 6 months ago
- FPGA-based hardware acceleration for dropout-based Bayesian Neural Networks.☆20Updated last year
- [TECS'23] A project on the co-design of Accelerators and CNNs.☆17Updated last year
- ☆18Updated 2 years ago
- ☆27Updated 4 years ago
- 2020 xilinx summer school☆16Updated 4 years ago
- [FPGA-2022] N3H-Core: Neuron-designed Neural Network Accelerator via FPGA-based Heterogeneous Computing Cores☆12Updated 2 years ago
- Edge-MoE: Memory-Efficient Multi-Task Vision Transformer Architecture with Task-level Sparsity via Mixture-of-Experts☆82Updated 4 months ago
- ☆20Updated last year
- ☆35Updated this week
- Fast Emulation of Approximate DNN Accelerators in PyTorch☆14Updated 6 months ago
- PolyLUT is the first quantized neural network training methodology that maps a neuron to a LUT while using multivariate polynomial functi…☆37Updated 7 months ago
- Provides the code for the paper "EBPC: Extended Bit-Plane Compression for Deep Neural Network Inference and Training Accelerators" by Luk…☆17Updated 4 years ago
- [TCAD'23] AccelTran: A Sparsity-Aware Accelerator for Transformers☆30Updated 9 months ago
- ☆16Updated last year
- Torch2Chip (MLSys, 2024)☆49Updated 3 weeks ago
- A collection of tutorials for the fpgaConvNet framework.☆28Updated last month
- SSR: Spatial Sequential Hybrid Architecture for Latency Throughput Tradeoff in Transformer Acceleration (Full Paper Accepted in FPGA'24)☆23Updated last month
- Adaptive floating-point based numerical format for resilient deep learning☆14Updated 2 years ago