[ICML24] Pruner-Zero: Evolving Symbolic Pruning Metric from scratch for LLMs
☆99Nov 25, 2024Updated last year
Alternatives and similar repositories for Pruner-Zero
Users that are interested in Pruner-Zero are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- D^2-MoE: Delta Decompression for MoE-based LLMs Compression☆82Mar 25, 2025Updated last year
- Are gradient information useful for pruning of LLMs?☆47Aug 23, 2025Updated 8 months ago
- [NeurIPS 2024] Search for Efficient LLMs☆16Jan 16, 2025Updated last year
- Awesome list for LLM pruning.☆291Oct 11, 2025Updated 6 months ago
- ☆12Oct 9, 2023Updated 2 years ago
- Deploy to Railway using AI coding agents - Free Credits Offer • AdUse Claude Code, Codex, OpenCode, and more. Autonomous software development now has the infrastructure to match with Railway.
- ☆57Jun 10, 2024Updated last year
- A block pruning framework for LLMs.☆28May 17, 2025Updated 11 months ago
- ☆41Nov 22, 2025Updated 5 months ago
- SpInfer: Leveraging Low-Level Sparsity for Efficient Large Language Model Inference on GPUs☆64Mar 25, 2025Updated last year
- Implementation of PGONAS for CVPR22W and RD-NAS for ICASSP23☆23Apr 25, 2023Updated 3 years ago
- [AAAI 2024] Fluctuation-based Adaptive Structured Pruning for Large Language Models☆75Jan 6, 2024Updated 2 years ago
- BESA is a differentiable weight pruning technique for large language models.☆17Mar 4, 2024Updated 2 years ago
- ☆23Nov 26, 2024Updated last year
- ☆63Dec 15, 2024Updated last year
- Serverless GPU API endpoints on Runpod - Get Bonus Credits • AdSkip the infrastructure headaches. Auto-scaling, pay-as-you-go, no-ops approach lets you focus on innovating your application.
- [ICML2024] DetKDS: Knowledge Distillation Search for Object Detectors☆19Jul 11, 2024Updated last year
- [ICRA 2025]Robust Self-Reconfiguration for Fault-Tolerant Control of Modular Aerial Robot Systems☆26Jun 9, 2025Updated 10 months ago
- This is the official repo for "Differentiable Model Scaling using Differentiable Topk"☆12May 16, 2024Updated last year
- ☆12Sep 1, 2023Updated 2 years ago
- A simple and effective LLM pruning approach.☆863Aug 9, 2024Updated last year