Efficient-ML / Awesome-Efficient-AIGCLinks
A list of papers, docs, codes about efficient AIGC. This repo is aimed to provide the info for efficient AIGC research, including language and vision, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.
β201Updated 9 months ago
Alternatives and similar repositories for Awesome-Efficient-AIGC
Users that are interested in Awesome-Efficient-AIGC are comparing it to the libraries listed below
Sorting:
- [NeurIPS 2024 Oralπ₯] DuQuant: Distributing Outliers via Dual Transformation Makes Stronger Quantized LLMs.β176Updated last year
- Code Repository of Evaluating Quantized Large Language Modelsβ135Updated last year
- Implementation of Post-training Quantization on Diffusion Models (CVPR 2023)β139Updated 2 years ago
- [ICML 2025] Official PyTorch implementation of "FlatQuant: Flatness Matters for LLM Quantization"β190Updated last month
- Code for the NeurIPS 2022 paper "Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning".β129Updated 2 years ago
- Code implementation of GPTAQ (https://arxiv.org/abs/2504.02692)β75Updated 3 months ago
- [ICLR'25] ViDiT-Q: Efficient and Accurate Quantization of Diffusion Transformers for Image and Video Generationβ132Updated 7 months ago
- The code repository of "MBQ: Modality-Balanced Quantization for Large Vision-Language Models"β66Updated 8 months ago
- The official PyTorch implementation of the ICLR2022 paper, QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quanβ¦β124Updated last month
- [CVPR 2024 Highlight & TPAMI 2025] This is the official PyTorch implementation of "TFMQ-DM: Temporal Feature Maintenance Quantization forβ¦β109Updated last month
- An algorithm for weight-activation quantization (W4A4, W4A8) of LLMs, supporting both static and dynamic quantizationβ164Updated 2 weeks ago
- β26Updated 11 months ago
- PyTorch implementation of PTQ4DiT https://arxiv.org/abs/2405.16005β40Updated last year
- Efficient Mixture of Experts for LLM Paper Listβ143Updated last month
- [ICLR2025]: OSTQuant: Refining Large Language Model Quantization with Orthogonal and Scaling Transformations for Better Distribution Fittβ¦β82Updated 7 months ago
- [CVPR 2025] Q-DiT: Accurate Post-Training Quantization for Diffusion Transformersβ72Updated last year
- A collection of research papers on low-precision training methodsβ43Updated 6 months ago
- [ICLR'25] ARB-LLM: Alternating Refined Binarizations for Large Language Modelsβ27Updated 3 months ago
- Awesome list for LLM quantizationβ353Updated last month
- Awesome list for LLM pruning.β273Updated last month
- The official implementation of the ICML 2023 paper OFQ-ViTβ33Updated 2 years ago
- Curated list of methods that focuses on improving the efficiency of diffusion modelsβ44Updated last year
- [ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMsβ121Updated 4 months ago
- [ICCV 2025] QuEST: Efficient Finetuning for Low-bit Diffusion Modelsβ55Updated 4 months ago
- This repository contains integer operators on GPUs for PyTorch.β222Updated 2 years ago
- [CVPR 2023] PD-Quant: Post-Training Quantization Based on Prediction Difference Metricβ60Updated 2 years ago
- Awesome LLM pruning papers all-in-one repository with integrating all useful resources and insights.β132Updated 3 months ago
- Code for the AAAI 2024 Oral paper "OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Modelβ¦β67Updated last year
- [COLM 2025] Official PyTorch implementation of "Quantization Hurts Reasoning? An Empirical Study on Quantized Reasoning Models"β57Updated 4 months ago
- [IJCAI 2022] FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformerβ354Updated 2 years ago