pvti / Awesome-Tensor-DecompositionLinks
π A curated list of tensor decomposition resources for model compression.
β77Updated last week
Alternatives and similar repositories for Awesome-Tensor-Decomposition
Users that are interested in Awesome-Tensor-Decomposition are comparing it to the libraries listed below
Sorting:
- A thoroughly investigated survey for tensorial neural networks.β137Updated 6 months ago
- β44Updated last year
- [ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Diβ¦β66Updated 10 months ago
- Collect optimizer related papers, data, repositoriesβ94Updated 8 months ago
- Official code implementation for 2025 ICLR accepted paper "Dobi-SVD : Differentiable SVD for LLM Compression and Some New Perspectives"β36Updated 4 months ago
- β13Updated 3 years ago
- β266Updated 11 months ago
- Efficient Expert Pruning for Sparse Mixture-of-Experts Language Models: Enhancing Performance and Reducing Inference Costsβ18Updated 7 months ago
- Fast Hadamard transform in CUDA, with a PyTorch interfaceβ215Updated last year
- This repo contains the code for studying the interplay between quantization and sparsity methodsβ22Updated 5 months ago
- Compressible Dynamics in Deep Overparameterized Low-Rank Learning & Adaptation (ICML'24 Oral)β13Updated last year
- Neural Tangent Kernel Papersβ115Updated 6 months ago
- [NAACL 24 Oral] LoRETTA: Low-Rank Economic Tensor-Train Adaptation for Ultra-Low-Parameter Fine-Tuning of Large Language Modelsβ36Updated 7 months ago
- Github Repo for OATS: Outlier-Aware Pruning through Sparse and Low Rank Decompositionβ13Updated 3 months ago
- Efficient tensor decomposition-based filter pruningβ17Updated 2 weeks ago
- β210Updated 2 years ago
- β49Updated last year
- (NeurIPS 2024) QuanTA: Efficient High-Rank Fine-Tuning of LLMs with Quantum-Informed Tensor Adaptationβ30Updated 8 months ago
- Torch2Chip (MLSys, 2024)β53Updated 4 months ago
- β26Updated last week
- Official Pytorch Implementation of "Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity"β73Updated last month
- [ICML 2024] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".β109Updated last month
- [ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMsβ111Updated last month
- β51Updated last year
- [ICLR2025]: OSTQuant: Refining Large Language Model Quantization with Orthogonal and Scaling Transformations for Better Distribution Fittβ¦β72Updated 4 months ago
- Awasome Papers and Resources in Deep Neural Network Pruning with Source Code.β161Updated 11 months ago
- Code for the AAAI 2024 Oral paper "OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Modelβ¦β64Updated last year
- Code Repository of Evaluating Quantized Large Language Modelsβ129Updated 11 months ago
- β59Updated last year
- Official implementation for Yuan & Liu & Zhong et al., KV Cache Compression, But What Must We Give in Return? A Comprehensive Benchmark oβ¦β81Updated 5 months ago