htqin / awesome-efficient-aigc
A list of papers, docs, codes about efficient AIGC. This repo is aimed to provide the info for efficient AIGC research, including language and vision, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.
☆142Updated 4 months ago
Related projects: ⓘ
- Awesome LLM pruning papers all-in-one repository with integrating all useful resources and insights.☆31Updated last month
- Code Repository of Evaluating Quantized Large Language Models☆89Updated last week
- [CVPR 2024 Highlight] This is the official PyTorch implementation of "TFMQ-DM: Temporal Feature Maintenance Quantization for Diffusion Mo…☆53Updated last month
- Implementation of Post-training Quantization on Diffusion Models (CVPR 2023)☆121Updated last year
- Awesome list for LLM pruning.☆130Updated 3 weeks ago
- [ICML 2024 Oral] This project is the official implementation of our Accurate LoRA-Finetuning Quantization of LLMs via Information Retenti…☆55Updated 5 months ago
- This is the official PyTorch implementation of "LLMC: Benchmarking Large Language Model Quantization with a Versatile Compression Toolkit…☆227Updated this week
- [ICML 2023] This project is the official implementation of our accepted ICML 2023 paper BiBench: Benchmarking and Analyzing Network Binar…☆54Updated 6 months ago
- QuEST: Efficient Finetuning for Low-bit Diffusion Models☆27Updated last month
- The official PyTorch implementation of the ICLR2022 paper, QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quan…☆107Updated last year
- The official implementation of PTQD: Accurate Post-Training Quantization for Diffusion Models☆85Updated 6 months ago
- [ICLR 2024 Spotlight] This is the official PyTorch implementation of "EfficientDM: Efficient Quantization-Aware Fine-Tuning of Low-Bit Di…☆49Updated 3 months ago
- Code for the NeurIPS 2022 paper "Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning".☆95Updated last year
- ViDiT-Q: Efficient and Accurate Quantization of Diffusion Transformers for Image and Video Generation☆23Updated 3 weeks ago
- This project is the official implementation of our accepted ICLR 2022 paper BiBERT: Accurate Fully Binarized BERT.☆81Updated last year
- Post-Training Quantization for Vision transformers.☆176Updated 2 years ago
- List of papers related to Vision Transformers quantization and hardware acceleration in recent AI conferences and journals.☆47Updated 3 months ago
- The official implementation of the NeurIPS 2022 paper Q-ViT.☆77Updated last year
- This repository contains integer operators on GPUs for PyTorch.☆172Updated 11 months ago
- [ACL 2024] A novel QAT with Self-Distillation framework to enhance ultra low-bit LLMs.☆69Updated 4 months ago
- Code repo for the paper "SpinQuant LLM quantization with learned rotations"☆79Updated this week
- The official PyTorch implementation of the NeurIPS2022 (spotlight) paper, Outlier Suppression: Pushing the Limit of Low-bit Transformer L…☆46Updated last year
- Official implementation of the EMNLP23 paper: Outlier Suppression+: Accurate quantization of large language models by equivalent and opti…☆38Updated 10 months ago
- [IJCAI 2022] FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformer☆301Updated last year
- Awesome list for LLM quantization☆84Updated 2 weeks ago
- Awasome Papers and Resources in Deep Neural Network Pruning with Source Code.☆119Updated 3 weeks ago
- The official implementation of the ICML 2023 paper OFQ-ViT☆27Updated 11 months ago
- Curated list of methods that focuses on improving the efficiency of diffusion models☆26Updated 2 months ago
- [NeurIPS 2022] A Fast Post-Training Pruning Framework for Transformers☆160Updated last year
- QAQ: Quality Adaptive Quantization for LLM KV Cache☆42Updated 5 months ago