Network acceleration methods
☆178Jun 19, 2021Updated 4 years ago
Alternatives and similar repositories for Network-Speed-and-Compression
Users that are interested in Network-Speed-and-Compression are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Papers for deep neural network compression and acceleration☆401Jun 21, 2021Updated 4 years ago
- papers about model compression☆166Feb 10, 2023Updated 3 years ago
- collection of works aiming at reducing model sizes or the ASIC/FPGA accelerator for machine learning☆568Feb 3, 2024Updated 2 years ago
- ☆34Feb 28, 2017Updated 9 years ago
- Collection of recent methods on (deep) neural network compression and acceleration.☆954Apr 4, 2025Updated last year
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- ☆670Aug 25, 2021Updated 4 years ago
- a list of awesome papers on deep model ompression and acceleration☆350Jun 19, 2021Updated 4 years ago
- Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)☆1,089May 2, 2024Updated last year
- Implements quantized distillation. Code for our paper "Model compression via distillation and quantization"☆336Jul 25, 2024Updated last year
- LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks☆245Aug 30, 2022Updated 3 years ago
- Summary, Code for Deep Neural Network Quantization☆558Jun 14, 2025Updated 10 months ago
- A list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures,…☆857Jun 19, 2021Updated 4 years ago
- Implementation for Trained Ternary Network.☆108Jan 13, 2017Updated 9 years ago
- ☆13Sep 12, 2017Updated 8 years ago
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- Compress neural network with pruning and quantization using TensorFlow.☆106Dec 19, 2018Updated 7 years ago
- Reducing the size of convolutional neural networks☆113Nov 28, 2017Updated 8 years ago
- A curated list of resources about few-shot and one-shot learning☆284Apr 22, 2019Updated 6 years ago
- PyTorch Implementation of [1611.06440] Pruning Convolutional Neural Networks for Resource Efficient Inference☆886Jul 12, 2019Updated 6 years ago
- PyTorch Model Compression☆234Jan 27, 2023Updated 3 years ago
- Hands-on Tutorial on Automated Deep Learning☆148Aug 27, 2020Updated 5 years ago
- A curated list of neural network pruning resources.☆2,492Apr 4, 2024Updated 2 years ago
- online hard examples mining support for Faster R-CNN end to end.☆11Aug 22, 2017Updated 8 years ago
- micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantiz…☆2,269May 6, 2025Updated 11 months ago
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- Codes for accepted paper "Cooperative Pruning in Cross-Domain Deep Neural Network Compression" in IJCAI 2019.☆12Aug 15, 2019Updated 6 years ago
- Caffe/Neon prototxt training file for our Neurocomputing2017 work: Fuzzy Quantitative Deep Compression Network☆12May 30, 2018Updated 7 years ago
- Code for “Discrimination-aware-Channel-Pruning-for-Deep-Neural-Networks”☆183Oct 29, 2020Updated 5 years ago
- An asynchronous pytorch Dataloader for general neural network pipeline accelaration.☆53Jan 7, 2024Updated 2 years ago
- Pytorch implementation for FAT: learning low-bitwidth parametric representation via frequency-aware transformation☆67May 2, 2021Updated 4 years ago
- Sparsifying Variational Dropout in Tensorflow☆22Nov 3, 2017Updated 8 years ago
- Code for: "And the bit goes down: Revisiting the quantization of neural networks"☆631Nov 9, 2020Updated 5 years ago
- Rethinking the Value of Network Pruning (Pytorch) (ICLR 2019)☆1,515Jun 7, 2020Updated 5 years ago
- Quantized Neural Network PACKage - mobile-optimized implementation of quantized neural network operators☆1,548Aug 28, 2019Updated 6 years ago
- Serverless GPU API endpoints on Runpod - Bonus Credits • AdSkip the infrastructure headaches. Auto-scaling, pay-as-you-go, no-ops approach lets you focus on innovating your application.
- Neural Network Quantization & Low-Bit Fixed Point Training For Hardware-Friendly Algorithm Design☆161Dec 18, 2020Updated 5 years ago
- Quantized Neural Networks - networks trained for inference at arbitrary low precision.☆147Nov 28, 2017Updated 8 years ago
- Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1☆1,064Nov 28, 2018Updated 7 years ago
- Implementation of ICLR 2017 paper "Loss-aware Binarization of Deep Networks"☆20Feb 24, 2019Updated 7 years ago
- This Toolkit is the fastest way to train YOLO-v2 with your own custom dataset from scratch including annotating part☆15Oct 9, 2020Updated 5 years ago
- Mayo: Auto-generation of hardware-friendly deep neural networks. Dynamic Channel Pruning: Feature Boosting and Suppression.☆111Dec 10, 2019Updated 6 years ago
- Model Scope in PyTorch (include Params, FLOPs, Madds).☆121May 28, 2019Updated 6 years ago