vantienpham / Awesome-Tensor-DecompositionLinks
π A curated list of tensor decomposition resources for model compression.
β105Updated last month
Alternatives and similar repositories for Awesome-Tensor-Decomposition
Users that are interested in Awesome-Tensor-Decomposition are comparing it to the libraries listed below
Sorting:
- [ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Diβ¦β70Updated last year
- Collect optimizer related papers, data, repositoriesβ99Updated last year
- Second-Order Fine-Tuning without Pain for LLMs: a Hessian Informed Zeroth-Order Optimizerβ23Updated last year
- β40Updated 2 months ago
- β13Updated 4 years ago
- Efficient Expert Pruning for Sparse Mixture-of-Experts Language Models: Enhancing Performance and Reducing Inference Costsβ23Updated 3 months ago
- [ICMLβ24] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".β123Updated 7 months ago
- Neural Tangent Kernel Papersβ121Updated last year
- Official code implementation for 2025 ICLR accepted paper "Dobi-SVD : Differentiable SVD for LLM Compression and Some New Perspectives"β50Updated 3 months ago
- β291Updated last year
- This repo contains the code for studying the interplay between quantization and sparsity methodsβ26Updated 11 months ago
- β44Updated 2 years ago
- This repository contains low-bit quantization papers from 2020 to 2025 on top conference.β95Updated 4 months ago
- (NeurIPS 2024) QuanTA: Efficient High-Rank Fine-Tuning of LLMs with Quantum-Informed Tensor Adaptationβ35Updated 2 months ago
- Github Repo for OATS: Outlier-Aware Pruning through Sparse and Low Rank Decompositionβ17Updated 9 months ago
- [NAACL 24 Oral] LoRETTA: Low-Rank Economic Tensor-Train Adaptation for Ultra-Low-Parameter Fine-Tuning of Large Language Modelsβ39Updated last year
- Reading list for research topics in state-space modelsβ345Updated 8 months ago
- Fast Hadamard transform in CUDA, with a PyTorch interfaceβ281Updated 3 months ago
- Official implementation for the paper "Understanding Hyperdimensional Computing for Parallel Single-Pass Learning"β23Updated 2 years ago
- Compressible Dynamics in Deep Overparameterized Low-Rank Learning & Adaptation (ICML'24 Oral)β13Updated last year
- Welcome to the 'In Context Learning Theory' Reading Groupβ30Updated last year
- SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretraining (NeurIPS 2024)β39Updated last year
- β222Updated 2 years ago
- Official Pytorch Implementation of "Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity"β80Updated 7 months ago
- [ICLR25] STBLLM: Breaking the 1-Bit Barrier with Structured Binary LLMsβ18Updated 8 months ago
- β56Updated last year
- summer school materialsβ46Updated 2 years ago
- [ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMsβ123Updated 7 months ago
- Apply CP, Tucker, TT/TR, HT to compress neural networks. Train from scratch.β17Updated 5 years ago
- [COLM 2025] Official PyTorch implementation of "Quantization Hurts Reasoning? An Empirical Study on Quantized Reasoning Models"β67Updated 7 months ago