This repository containts the pytorch scripts to train mixed-precision networks for microcontroller deployment, based on the memory contraints of the target device.
☆51May 9, 2024Updated last year
Alternatives and similar repositories for training-mixed-precision-quantized-networks
Users that are interested in training-mixed-precision-quantized-networks are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- source code of the paper: Robust Quantization: One Model to Rule Them All☆41Mar 24, 2023Updated 3 years ago
- Conditional channel- and precision-pruning on neural networks☆71Mar 4, 2020Updated 6 years ago
- ProxQuant: Quantized Neural Networks via Proximal Operators☆30Feb 19, 2019Updated 7 years ago
- [CVPR'20] ZeroQ Mixed-Precision implementation (unofficial): A Novel Zero Shot Quantization Framework☆14Dec 16, 2020Updated 5 years ago
- ☆45Jan 17, 2020Updated 6 years ago
- Managed Database hosting by DigitalOcean • AdPostgreSQL, MySQL, MongoDB, Kafka, Valkey, and OpenSearch available. Automatically scale up storage and focus on building your apps.
- [CVPR 2019, Oral] HAQ: Hardware-Aware Automated Quantization with Mixed Precision☆405Feb 26, 2021Updated 5 years ago
- ☆213Nov 23, 2018Updated 7 years ago
- DNN quantization with outlier channel splitting (ICML'19)☆114Mar 21, 2020Updated 6 years ago
- Codes for AAAI2019 paper: Deep Neural Network Quantization via Layer-Wise Optimization using Limited Training Data☆41Jan 22, 2019Updated 7 years ago
- [CVPR'20] ZeroQ: A Novel Zero Shot Quantization Framework☆282Dec 8, 2023Updated 2 years ago
- Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming☆98Jun 10, 2021Updated 4 years ago
- Training Quantized Neural Networks with a Full-precision Auxiliary Module☆13Jun 19, 2020Updated 5 years ago
- LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks☆245Aug 30, 2022Updated 3 years ago
- PyTorch code for full quantization of DNN using BCGD☆14Jul 24, 2019Updated 6 years ago
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- Caffe implementation of Optimal-Ternary-Weights-Approximation in "Two-Step Quantization for Low-bit Neural Networks" (CVPR2018).☆15Sep 21, 2018Updated 7 years ago
- Caffe Implementation for Incremental network quantization☆191Jul 29, 2018Updated 7 years ago
- [NeurIPS'2019] Shupeng Gui, Haotao Wang, Haichuan Yang, Chen Yu, Zhangyang Wang, Ji Liu, “Model Compression with Adversarial Robustness: …☆48Dec 30, 2021Updated 4 years ago
- Simulator for BitFusion☆101Aug 6, 2020Updated 5 years ago
- PyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.☆264Oct 3, 2023Updated 2 years ago
- A PyTorch implementation of "Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights"☆167Mar 8, 2020Updated 6 years ago
- NEural Minimizer for pytOrch☆47Jul 25, 2024Updated last year
- PyTorch implementation for the APoT quantization (ICLR 2020)☆286Dec 11, 2024Updated last year
- The goal of this design is to use the PYNQ-Z2 development board to design a general convolution neural network accelerator. And through r…☆11Sep 30, 2020Updated 5 years ago
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- Code for reproducing the results from "CrAM: A Compression-Aware Minimizer" accepted at ICLR 2023☆10Mar 1, 2023Updated 3 years ago
- Bi-Real-Net Model in Pytorch https://arxiv.org/abs/1808.00278 + pre-trained fp model weights☆18Sep 17, 2019Updated 6 years ago
- CMix-NN: Mixed Low-Precision CNN Library for Memory-Constrained Edge Devices☆49Mar 19, 2020Updated 6 years ago
- A generic code base for neural network pruning, especially for pruning at initialization.☆32Sep 3, 2022Updated 3 years ago
- Deep Compression for PyTorch Model Deployment on Microcontrollers☆19Mar 26, 2021Updated 5 years ago
- ☆48Jul 28, 2020Updated 5 years ago
- ☆19Mar 16, 2022Updated 4 years ago
- Proximal Mean-field for Neural Network Quantization☆21Apr 9, 2020Updated 6 years ago
- ☆15Jan 8, 2020Updated 6 years ago
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- Implementation of ICLR 2018 paper "Loss-aware Weight Quantization of Deep Networks"☆27Oct 24, 2019Updated 6 years ago
- A tool to deploy Deep Neural Networks on PULP-based SoC's☆92Aug 4, 2025Updated 8 months ago
- PyTorch implementation of "Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference"☆56Aug 17, 2019Updated 6 years ago
- All about acceleration and compression of Deep Neural Networks☆33Nov 5, 2019Updated 6 years ago
- ☆72Feb 16, 2023Updated 3 years ago
- Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming☆36Jun 29, 2023Updated 2 years ago
- ☆172Mar 9, 2023Updated 3 years ago