pvti / Awesome-Tensor-DecompositionLinks
😎 A curated list of tensor decomposition resources for model compression.
☆73Updated this week
Alternatives and similar repositories for Awesome-Tensor-Decomposition
Users that are interested in Awesome-Tensor-Decomposition are comparing it to the libraries listed below
Sorting:
- A thoroughly investigated survey for tensorial neural networks.☆136Updated 6 months ago
- [ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Di…☆62Updated 9 months ago
- Official code implementation for 2025 ICLR accepted paper "Dobi-SVD : Differentiable SVD for LLM Compression and Some New Perspectives"☆34Updated 3 months ago
- Efficient Expert Pruning for Sparse Mixture-of-Experts Language Models: Enhancing Performance and Reducing Inference Costs☆18Updated 7 months ago
- [NAACL 24 Oral] LoRETTA: Low-Rank Economic Tensor-Train Adaptation for Ultra-Low-Parameter Fine-Tuning of Large Language Models☆35Updated 6 months ago
- [ICML 2024] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".☆105Updated last week
- Collect optimizer related papers, data, repositories☆92Updated 8 months ago
- Compressible Dynamics in Deep Overparameterized Low-Rank Learning & Adaptation (ICML'24 Oral)☆13Updated 11 months ago
- ☆13Updated 3 years ago
- Github Repo for OATS: Outlier-Aware Pruning through Sparse and Low Rank Decomposition☆13Updated 2 months ago
- ☆42Updated last year
- Fast Hadamard transform in CUDA, with a PyTorch interface☆201Updated last year
- This repo contains the code for studying the interplay between quantization and sparsity methods☆21Updated 4 months ago
- ☆260Updated 10 months ago
- Reading list for research topics in state-space models☆306Updated last month
- ☆52Updated 11 months ago
- Official code for the paper "Examining Post-Training Quantization for Mixture-of-Experts: A Benchmark"☆20Updated 2 weeks ago
- Official Repo for SparseLLM: Global Pruning of LLMs (NeurIPS 2024)☆64Updated 3 months ago
- Efficient tensor decomposition-based filter pruning☆16Updated last year
- ☆22Updated last month
- A library for calculating the FLOPs in the forward() process based on torch.fx☆118Updated 3 months ago
- Activation-aware Singular Value Decomposition for Compressing Large Language Models☆74Updated 8 months ago
- LLM Inference with Microscaling Format☆24Updated 8 months ago
- Awesome LLM pruning papers all-in-one repository with integrating all useful resources and insights.☆96Updated 7 months ago
- Neural Tangent Kernel Papers☆115Updated 6 months ago
- ☆24Updated 2 months ago
- ☆58Updated last year
- ☆50Updated 9 months ago
- SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretraining (NeurIPS 2024)☆32Updated 8 months ago
- [ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMs☆111Updated last week