PannenetsF / TQTLinks
TQT's pytorch implementation.
☆21Updated 3 years ago
Alternatives and similar repositories for TQT
Users that are interested in TQT are comparing it to the libraries listed below
Sorting:
- BitSplit Post-trining Quantization☆50Updated 3 years ago
- Neural Network Quantization With Fractional Bit-widths☆11Updated 4 years ago
- A Out-of-box PyTorch Scaffold for Neural Network Quantization-Aware-Training (QAT) Research. Website: https://github.com/zhutmost/neuralz…☆25Updated 2 years ago
- ☆28Updated 3 years ago
- Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming☆98Updated 4 years ago
- DAC System Design Contest 2020☆29Updated 5 years ago
- Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming☆35Updated 2 years ago
- BSQ: Exploring Bit-Level Sparsity for Mixed-Precision Neural Network Quantization (ICLR 2021)☆41Updated 4 years ago
- This repository containts the pytorch scripts to train mixed-precision networks for microcontroller deployment, based on the memory contr…☆50Updated last year
- ☆23Updated 4 years ago
- Post-training sparsity-aware quantization☆34Updated 2 years ago
- ☆19Updated 4 years ago
- Static Block Floating Point Quantization for CNN☆36Updated 4 years ago
- [CVPR'20] ZeroQ Mixed-Precision implementation (unofficial): A Novel Zero Shot Quantization Framework☆14Updated 4 years ago
- [ICLR 2022 Oral] F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization☆94Updated 3 years ago
- ☆35Updated 6 years ago
- Training with Block Minifloat number representation☆16Updated 4 years ago
- Code for our paper at ECCV 2020: Post-Training Piecewise Linear Quantization for Deep Neural Networks☆68Updated 3 years ago
- ☆20Updated 3 years ago
- This is an implementation of YOLO using LSQ network quantization method.☆22Updated 3 years ago
- ☆57Updated 4 years ago
- ☆76Updated 3 years ago
- ☆36Updated 6 years ago
- ☆19Updated 3 years ago
- HW/SW co-design of sentence-level energy optimizations for latency-aware multi-task NLP inference☆52Updated last year
- Implementation of "DeepShift: Towards Multiplication-Less Neural Networks" https://arxiv.org/abs/1905.13298☆112Updated 3 years ago
- [TCAD 2021] Block Convolution: Towards Memory-Efficient Inference of Large-Scale CNNs on FPGA☆17Updated 3 years ago
- PyTorch implementation of EdMIPS: https://arxiv.org/pdf/2004.05795.pdf☆60Updated 5 years ago
- ☆12Updated 3 years ago
- An open-sourced PyTorch library for developing energy efficient multiplication-less models and applications.☆13Updated 8 months ago