AldrichZeng / Graduation-DesignLinks
基于剪枝的神经网络压缩与加速
☆23Updated 6 years ago
Alternatives and similar repositories for Graduation-Design
Users that are interested in Graduation-Design are comparing it to the libraries listed below
Sorting:
- 模型压缩demo(剪枝、量化、知识蒸馏)☆77Updated 5 years ago
- In this repository using the sparse training, group channel pruning and knowledge distilling for YOLOV4,☆31Updated 2 years ago
- B站Efficient-Neural-Network学习分享的配套代码☆300Updated 3 years ago
- Unofficial Pytorch implementation of Deep Compression in CIFAR10☆35Updated 3 years ago
- 基于提前退出部分样本原理而实现的带分支网络(supported by chainer)☆45Updated 6 years ago
- ☆36Updated 6 years ago
- PyTorch implementation of "Pruning Filters For Efficient ConvNets"☆151Updated last year
- Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks☆380Updated 5 years ago
- Base to channel pruned to ResNet18 model☆145Updated 2 years ago
- 对yolov4进行通道剪枝☆15Updated 3 years ago
- A reproduction of PRUNING FILTERS FOR EFFICIENT CONVNETS☆28Updated 5 years ago
- ☆17Updated 6 years ago
- An implementation of ResNet with mixup and cutout regularizations and soft filter pruning.☆16Updated 4 years ago
- An 8bit automated quantization conversion tool for the pytorch (Post-training quantization based on KL divergence)☆33Updated 5 years ago
- Pruned model: VGG & ResNet-50☆17Updated 6 years ago
- yolov3_tiny implement on tensoeflow for int8 quantization (tflite)☆29Updated 6 years ago
- OpenPose uses Pytorch for static quantization, saving, and loading of models☆88Updated 3 years ago
- About DNN compression and acceleration on Edge Devices.☆55Updated 4 years ago
- Learning both Weights and Connections for Efficient Neural Networks https://arxiv.org/abs/1506.02626☆18Updated 4 years ago
- Pytorch Implementationg of “Learning Efficient Convolutional Networks through Network Slimming”☆77Updated 6 years ago
- Compression of Deep Neural Networks LeNet-300-100 and LeNet-5 trained on MNIST and CIFAR-10 using Quantization, Knowledge Distillation & …☆20Updated 5 years ago
- base quantization methods including: QAT, PTQ, per_channel, per_tensor, dorefa, lsq, adaround, omse, Histogram, bias_correction.etc☆46Updated 2 years ago
- pytorch AutoSlim tools,支持三行代码对pytorch模型进行剪枝压缩☆39Updated 4 years ago
- Quantize,Pytorch,Vgg16,MobileNet☆42Updated 4 years ago
- Code and Results for Master Thesis Project on Fixed-point Quantization of Convolutional Neural Networks for Quantized Inference on Embedd…☆12Updated 4 years ago
- model compression and optimization for deployment for Pytorch, including knowledge distillation, quantization and pruning.(知识蒸馏,量化,剪枝)☆18Updated 9 months ago
- model compression based on pytorch (1、quantization: 8/4/2bits(dorefa)、ternary/binary value(twn/bnn/xnor-net);2、 pruning: normal、regular a…☆170Updated 4 years ago
- 用于 MobileNetV3 在自定义数据集上的量化,模型压缩90%而精度几乎不受影响,论文:HAQ: Hardware-Aware Automated Quantization with Mixed Precision☆18Updated 3 years ago
- my blog☆55Updated 5 years ago
- Model Compression 1. Pruning(BN Pruning) 2. Knowledge Distillation (Hinton) 3. Quantization (MNN) 4. Deployment (MNN)☆79Updated 4 years ago