BESA is a differentiable weight pruning technique for large language models.
☆17Mar 4, 2024Updated last year
Alternatives and similar repositories for LLMPrune-BESA
Users that are interested in LLMPrune-BESA are comparing it to the libraries listed below
Sorting:
- super-resolution; post-training quantization; model compression☆14Nov 10, 2023Updated 2 years ago
- Pytorch implementation of our paper accepted by ICML 2023 -- "Bi-directional Masks for Efficient N:M Sparse Training"☆12Jun 7, 2023Updated 2 years ago
- [ACL'22] Training-free Neural Architecture Search for RNNs and Transformers☆14May 26, 2024Updated last year
- ☆56Jun 10, 2024Updated last year
- ☆16Dec 9, 2023Updated 2 years ago
- Pytorch最小实践☆17Jun 27, 2022Updated 3 years ago
- ☆11Feb 5, 2026Updated 3 weeks ago
- Fork of Flame repo for training of some new stuff in development☆19Feb 20, 2026Updated last week
- ☆17Jul 10, 2022Updated 3 years ago
- [ICML2025] KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference☆26Jan 27, 2026Updated last month
- ☆42Apr 23, 2024Updated last year
- Pytorch implementation of our paper accepted by ECCV 2022-- Fine-grained Data Distribution Alignment for Post-Training Quantization☆15Sep 13, 2022Updated 3 years ago
- [NeurIPS 2023] Token-Scaled Logit Distillation for Ternary Weight Generative Language Models☆18Dec 6, 2023Updated 2 years ago
- ☆18Nov 6, 2019Updated 6 years ago
- Are gradient information useful for pruning of LLMs?☆47Aug 23, 2025Updated 6 months ago
- FireQ: Fast INT4-FP8 Kernel and RoPE-aware Quantization for LLM Inference Acceleration☆20Jun 27, 2025Updated 8 months ago
- This repo contains the code for studying the interplay between quantization and sparsity methods☆26Feb 26, 2025Updated last year
- Structural Pruning for LLaMA☆54May 20, 2023Updated 2 years ago
- Repository for "Accelerating Neural Architecture Search using Performance Prediction" (ICLR Workshop 2018)☆18Mar 21, 2018Updated 7 years ago
- AFPQ code implementation☆23Nov 6, 2023Updated 2 years ago
- ☆26Dec 10, 2020Updated 5 years ago
- ☆28Dec 2, 2024Updated last year
- ☆29Jun 11, 2023Updated 2 years ago
- [AAAI 2024] Fluctuation-based Adaptive Structured Pruning for Large Language Models☆70Jan 6, 2024Updated 2 years ago
- ☆30Jul 22, 2024Updated last year
- [AAAI '23] PINAT: A Permutation INvariance Augmented Transformer for NAS Predictor☆31Feb 25, 2023Updated 3 years ago
- Pytorch implementation of "Oscillation-Reduced MXFP4 Training for Vision Transformers" on DeiT Model Pre-training☆36Jun 20, 2025Updated 8 months ago
- ☆25Dec 11, 2021Updated 4 years ago
- An innovative method expediting LLMs via streamlined semi-autoregressive generation and draft verification.☆26Apr 15, 2025Updated 10 months ago
- ☆29Nov 29, 2023Updated 2 years ago
- Vstream - Video Analytics pipeline with Hardware based accelerations (dev - stage)☆10Feb 2, 2024Updated 2 years ago
- [ICCV-2023] EMQ: Evolving Training-free Proxies for Automated Mixed Precision Quantization☆28Dec 6, 2023Updated 2 years ago
- [ICML 2024] Sparse Model Inversion: Efficient Inversion of Vision Transformers with Less Hallucination☆13Apr 29, 2025Updated 10 months ago
- Repository for the paper: Teaching VLMs to Localize Specific Objects from In-context Examples☆40Nov 27, 2024Updated last year
- [NeurIPS 2023] ShiftAddViT: Mixture of Multiplication Primitives Towards Efficient Vision Transformer☆30Dec 6, 2023Updated 2 years ago
- [ICLR 2022] "Learning Pruning-Friendly Networks via Frank-Wolfe: One-Shot, Any-Sparsity, and No Retraining" by Lu Miao*, Xiaolong Luo*, T…☆33Jan 20, 2022Updated 4 years ago
- Pytorch implementation of our paper accepted by TPAMI 2023 — Lottery Jackpots Exist in Pre-trained Models☆35Jun 19, 2023Updated 2 years ago
- Code for "Everybody Prune Now: Structured Pruning of LLMs with only Forward Passes"☆30Mar 28, 2024Updated last year
- Continual Resilient (CoRe) Optimizer for PyTorch☆11Jun 10, 2024Updated last year