Efficient-ML / Awesome-Model-Quantization
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.
☆2,026Updated 2 weeks ago
Alternatives and similar repositories for Awesome-Model-Quantization:
Users that are interested in Awesome-Model-Quantization are comparing it to the libraries listed below
- List of papers related to neural network quantization in recent AI conferences and journals.☆562Updated 3 months ago
- A curated list of neural network pruning resources.☆2,426Updated 11 months ago
- Awesome machine learning model compression research papers, quantization, tools, and learning material.☆507Updated 6 months ago
- Model Quantization Benchmark☆793Updated 2 months ago
- Collection of recent methods on (deep) neural network compression and acceleration.☆939Updated 3 months ago
- Summary, Code for Deep Neural Network Quantization☆546Updated 5 months ago
- [CVPR 2023] DepGraph: Towards Any Structural Pruning☆2,928Updated 2 weeks ago
- A simple network quantization demo using pytorch from scratch.☆521Updated last year
- Pytorch implementation of various Knowledge Distillation (KD) methods.☆1,672Updated 3 years ago
- micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantiz…☆2,230Updated 3 years ago
- [ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment☆1,902Updated last year
- Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.☆428Updated last year
- Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。☆2,559Updated last year
- Awesome LLM compression research papers and tools.☆1,427Updated this week
- A coding-free framework built on PyTorch for reproducible deep learning studies. PyTorch Ecosystem. 🏆25 knowledge distillation methods p…☆1,467Updated last week
- OpenMMLab Model Compression Toolbox and Benchmark.☆1,559Updated 9 months ago
- A curated list for Efficient Large Language Models☆1,547Updated last week
- ☆231Updated 7 months ago
- ☆668Updated 3 years ago
- A general and accurate MACs / FLOPs profiler for PyTorch models☆601Updated 10 months ago
- PPL Quantization Tool (PPQ) is a powerful offline neural network quantization tool.☆1,654Updated 11 months ago
- The official implementation of [CVPR2022] Decoupled Knowledge Distillation https://arxiv.org/abs/2203.08679 and [ICCV2023] DOT: A Distill…☆840Updated last year
- knowledge distillation papers☆752Updated 2 years ago
- [TMLR 2024] Efficient Large Language Models: A Survey☆1,117Updated 3 weeks ago
- [IJCAI 2022] FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformer☆327Updated last year
- Awesome Knowledge Distillation☆3,611Updated last week
- PyTorch implementation for the APoT quantization (ICLR 2020)☆271Updated 3 months ago
- [ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices☆440Updated last year
- Efficient computing methods developed by Huawei Noah's Ark Lab☆1,254Updated 4 months ago
- A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility☆1,921Updated last year