Efficient-ML / Awesome-Model-QuantizationLinks
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.
☆2,119Updated 3 months ago
Alternatives and similar repositories for Awesome-Model-Quantization
Users that are interested in Awesome-Model-Quantization are comparing it to the libraries listed below
Sorting:
- List of papers related to neural network quantization in recent AI conferences and journals.☆641Updated 2 months ago
- A curated list of neural network pruning resources.☆2,452Updated last year
- Model Quantization Benchmark☆806Updated last month
- Awesome machine learning model compression research papers, quantization, tools, and learning material.☆524Updated 8 months ago
- Summary, Code for Deep Neural Network Quantization☆548Updated 7 months ago
- Pytorch implementation of various Knowledge Distillation (KD) methods.☆1,698Updated 3 years ago
- Collection of recent methods on (deep) neural network compression and acceleration.☆950Updated 2 months ago
- Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.☆435Updated 2 years ago
- micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantiz…☆2,248Updated last month
- A simple network quantization demo using pytorch from scratch.☆532Updated last year
- A curated list for Efficient Large Language Models☆1,694Updated last month
- [CVPR 2023] DepGraph: Towards Any Structural Pruning; LLMs, Vision Foundation Models, etc.☆3,022Updated last month
- [IJCAI 2022] FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformer☆343Updated 2 years ago
- Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。☆2,598Updated 2 years ago
- PPL Quantization Tool (PPQ) is a powerful offline neural network quantization tool.☆1,690Updated last year
- Awesome LLM compression research papers and tools.☆1,539Updated last week
- [ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models☆1,417Updated 10 months ago
- [ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment☆1,918Updated last year
- [TMLR 2024] Efficient Large Language Models: A Survey☆1,165Updated this week
- ☆248Updated 9 months ago
- [ICLR 2020] Contrastive Representation Distillation (CRD), and benchmark of recent knowledge distillation methods☆2,339Updated last year
- A list of papers, docs, codes about efficient AIGC. This repo is aimed to provide the info for efficient AIGC research, including languag…☆182Updated 3 months ago
- PyTorch implementation for the APoT quantization (ICLR 2020)☆273Updated 5 months ago
- A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility☆1,940Updated 2 years ago
- OpenMMLab Model Compression Toolbox and Benchmark.☆1,600Updated 11 months ago
- Pytorch implementation of BRECQ, ICLR 2021☆273Updated 3 years ago
- ☆668Updated 3 years ago
- Rethinking the Value of Network Pruning (Pytorch) (ICLR 2019)☆1,514Updated 5 years ago
- AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.☆2,323Updated this week
- [CVPR 2019, Oral] HAQ: Hardware-Aware Automated Quantization with Mixed Precision☆385Updated 4 years ago