BBuf / model-compressionLinks
model compression based on pytorch (1、quantization: 8/4/2bits(dorefa)、ternary/binary value(twn/bnn/xnor-net);2、 pruning: normal、regular and group convolutional channel pruning;3、 group convolution structure;4、batch-normalization folding for binary value of feature(A))
☆170Updated 5 years ago
Alternatives and similar repositories for model-compression
Users that are interested in model-compression are comparing it to the libraries listed below
Sorting:
- 针对pytorch模型的自动化模型结构分析和修改工具集,包含自动分析模型结构的模型压缩算法库☆254Updated 2 years ago
- ONNX2Pytorch☆164Updated 4 years ago
- pytorch AutoSlim tools,支持三行代码对pytorch模型进行剪枝压缩☆39Updated 4 years ago
- A tutorial about how to build a TensorRT Engine from a PyTorch Model with the help of ONNX☆247Updated 4 years ago
- www.giantpandacv.com☆152Updated last year
- Pytorch-->onnx-->TensorRT; CUDA11, CUDNN8, TensorRT8☆211Updated last year
- Everything in Torch Fx☆345Updated last year
- ResRep: Lossless CNN Pruning via Decoupling Remembering and Forgetting (ICCV 2021)☆297Updated 2 years ago
- A nnie quantization aware training tool on pytorch.☆238Updated 4 years ago
- RM Operation can equivalently convert ResNet to VGG, which is better for pruning; and can help RepVGG perform better when the depth is la…☆211Updated 2 years ago
- Try out different pruning-approaches on lightweight Backbones.☆146Updated 5 years ago
- EasyQuant(EQ) is an efficient and simple post-training quantization method via effectively optimizing the scales of weights and activatio…☆405Updated 2 years ago
- Personal Pytorch toy script.☆67Updated 3 years ago
- A simple network quantization demo using pytorch from scratch.☆538Updated 2 years ago
- ☆100Updated 4 years ago
- ☆14Updated 4 years ago
- ☆338Updated 3 years ago
- An Improved One millisecond Mobile Backbone☆147Updated 3 years ago
- 对 YOLOv3 做模型剪枝(network slimming),对于 oxford hand 数据集(因项目需要),模型剪枝后的参数量减少 80%,Infer 的速度达到原来 2 倍,mAP 基本不变☆12Updated 6 years ago
- pytorch -> onnx -> caffe, pytorch to caffe, or other deep learning framework to onnx and onnx to caffe.☆164Updated 4 years ago
- Model Compression 1. Pruning(BN Pruning) 2. Knowledge Distillation (Hinton) 3. Quantization (MNN) 4. Deployment (MNN)☆79Updated 4 years ago
- tensorrt int8 量化yolov5 onnx模型☆185Updated 4 years ago
- MobileNetV3 based SSD-lite implementation in Pytorch☆100Updated 6 years ago
- A module to convert pytorch model to caffe model.☆67Updated 5 years ago
- ghostnet_cifar10☆116Updated 5 years ago
- NVIDIA TensorRT 加速推断教程!☆134Updated 4 years ago
- Learning Efficient Convolutional Networks through Network Slimming, In ICCV 2017.☆575Updated 6 years ago
- Pruning and quantization for SSD. Model compression.☆30Updated 4 years ago
- RepVGG TensorRT int8 量化,实测推理不到1ms一帧!☆62Updated 4 years ago
- ☆81Updated 4 years ago