analogdevicesinc / distiller
Fork of Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://nervanasystems.github.io/distiller
☆14Updated 6 months ago
Alternatives and similar repositories for distiller:
Users that are interested in distiller are comparing it to the libraries listed below
- PyTorch Quantization Aware Training Example☆130Updated 9 months ago
- Inference of quantization aware trained networks using TensorRT☆80Updated 2 years ago
- PyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.☆260Updated last year
- Count number of parameters / MACs / FLOPS for ONNX models.☆90Updated 3 months ago
- Scailable ONNX python tools☆96Updated 3 months ago
- On-the-fly Structured Pruning for PyTorch models. This library implements several attributions metrics and structured pruning utils for n…☆164Updated 4 years ago
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆33Updated 3 years ago
- TFLite model analyzer & memory optimizer☆122Updated last year
- ☆312Updated last year
- PyTorch Pruning Example☆49Updated 2 years ago
- Binarize convolutional neural networks using pytorch☆139Updated 2 years ago
- PyTorch Static Quantization Example☆38Updated 3 years ago
- PyTorch implementation for the APoT quantization (ICLR 2020)☆271Updated 2 months ago
- Repository to track the progress in model compression and acceleration☆105Updated 3 years ago
- A code generator from ONNX to PyTorch code☆135Updated 2 years ago
- ☆10Updated last year
- Pytorch implementation of BRECQ, ICLR 2021☆263Updated 3 years ago
- Roughly calculate FLOPs of a tflite model☆37Updated 3 years ago
- [ICLR 2022 Oral] F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization☆94Updated 2 years ago
- A parser, editor and profiler tool for ONNX models.☆416Updated last month
- Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming☆96Updated 3 years ago
- Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. Th…☆361Updated this week
- quantize aware training package for NCNN on pytorch☆70Updated 3 years ago
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆31Updated 4 months ago
- Code for our paper at ECCV 2020: Post-Training Piecewise Linear Quantization for Deep Neural Networks☆68Updated 3 years ago
- Batch Normalization Auto-fusion for PyTorch☆32Updated 4 years ago
- ☆224Updated 3 years ago
- [CVPR 2020] APQ: Joint Search for Network Architecture, Pruning and Quantization Policy☆157Updated 4 years ago
- ☆136Updated last year
- Neural Network Quantization & Low-Bit Fixed Point Training For Hardware-Friendly Algorithm Design☆160Updated 4 years ago