Efficient-ML / Awesome-Efficient-AIGCLinks
A list of papers, docs, codes about efficient AIGC. This repo is aimed to provide the info for efficient AIGC research, including language and vision, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.
β193Updated 8 months ago
Alternatives and similar repositories for Awesome-Efficient-AIGC
Users that are interested in Awesome-Efficient-AIGC are comparing it to the libraries listed below
Sorting:
- [NeurIPS 2024 Oralπ₯] DuQuant: Distributing Outliers via Dual Transformation Makes Stronger Quantized LLMs.β171Updated last year
- Code Repository of Evaluating Quantized Large Language Modelsβ132Updated last year
- Implementation of Post-training Quantization on Diffusion Models (CVPR 2023)β139Updated 2 years ago
- Awesome list for LLM pruning.β264Updated this week
- [ICLR'25] ViDiT-Q: Efficient and Accurate Quantization of Diffusion Transformers for Image and Video Generationβ124Updated 6 months ago
- The code repository of "MBQ: Modality-Balanced Quantization for Large Vision-Language Models"β59Updated 6 months ago
- An algorithm for weight-activation quantization (W4A4, W4A8) of LLMs, supporting both static and dynamic quantizationβ153Updated 4 months ago
- Awesome list for LLM quantizationβ314Updated this week
- PyTorch implementation of PTQ4DiT https://arxiv.org/abs/2405.16005β33Updated 11 months ago
- Code for the NeurIPS 2022 paper "Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning".β128Updated 2 years ago
- [CVPR 2024 Highlight & TPAMI 2025] This is the official PyTorch implementation of "TFMQ-DM: Temporal Feature Maintenance Quantization forβ¦β104Updated last week
- Code implementation of GPTAQ (https://arxiv.org/abs/2504.02692)β65Updated 2 months ago
- [ICML 2025] Official PyTorch implementation of "FlatQuant: Flatness Matters for LLM Quantization"β171Updated last week
- Awesome LLM pruning papers all-in-one repository with integrating all useful resources and insights.β123Updated 2 months ago
- The official PyTorch implementation of the ICLR2022 paper, QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quanβ¦β125Updated 2 weeks ago
- Awesome Pruning. β Curated Resources for Neural Network Pruning.β168Updated last year
- [CVPR 2025] Q-DiT: Accurate Post-Training Quantization for Diffusion Transformersβ66Updated last year
- Curated list of methods that focuses on improving the efficiency of diffusion modelsβ45Updated last year
- β277Updated last year
- List of papers related to Vision Transformers quantization and hardware acceleration in recent AI conferences and journals.β94Updated last year
- A collection of research papers on low-precision training methodsβ37Updated 5 months ago
- [ICML'24 Oral] APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and Inferenceβ45Updated last year
- β24Updated 10 months ago
- [ICLR'25] ARB-LLM: Alternating Refined Binarizations for Large Language Modelsβ27Updated 2 months ago
- Post-Training Quantization for Vision transformers.β228Updated 3 years ago
- The official implementation of PTQD: Accurate Post-Training Quantization for Diffusion Modelsβ100Updated last year
- [ICLR 2024 Spotlight] This is the official PyTorch implementation of "EfficientDM: Efficient Quantization-Aware Fine-Tuning of Low-Bit Diβ¦β66Updated last year
- [ICML 2023] This project is the official implementation of our accepted ICML 2023 paper BiBench: Benchmarking and Analyzing Network Binarβ¦β55Updated last year
- β14Updated 6 months ago
- List of papers related to neural network quantization in recent AI conferences and journals.β731Updated 6 months ago