zimingyy / SubZero
☆11Updated 5 months ago
Alternatives and similar repositories for SubZero
Users that are interested in SubZero are comparing it to the libraries listed below
Sorting:
- Second-Order Fine-Tuning without Pain for LLMs: a Hessian Informed Zeroth-Order Optimizer☆14Updated 3 months ago
- [ICML 2024] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".☆102Updated 10 months ago
- [ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Di…☆57Updated 7 months ago
- [NeurIPS 2024 Oral🔥] DuQuant: Distributing Outliers via Dual Transformation Makes Stronger Quantized LLMs.☆159Updated 7 months ago
- Lossless Training Speed Up by Unbiased Dynamic Data Pruning☆333Updated 7 months ago
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆199Updated 5 months ago
- Implementation of the 2023 CVPR Award Candidate: On Distillation of Guided Diffusion Models☆48Updated last year
- Github Repo for OATS: Outlier-Aware Pruning through Sparse and Low Rank Decomposition☆12Updated 3 weeks ago
- The official PyTorch implementation of the ICLR2022 paper, QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quan…☆121Updated last year
- ☆53Updated 5 months ago
- The code repository of "MBQ: Modality-Balanced Quantization for Large Vision-Language Models"☆38Updated last month
- Code for the paper: Why Transformers Need Adam: A Hessian Perspective☆57Updated 2 months ago
- PyTorch implementation of PTQ4DiT https://arxiv.org/abs/2405.16005☆30Updated 6 months ago
- ICLR 2024, Towards Lossless Dataset Distillation via Difficulty-Aligned Trajectory Matching☆102Updated 11 months ago
- [CVPR 2024] On the Diversity and Realism of Distilled Dataset: An Efficient Dataset Distillation Paradigm☆69Updated 2 months ago
- A list of papers, docs, codes about efficient AIGC. This repo is aimed to provide the info for efficient AIGC research, including languag…☆180Updated 3 months ago
- ☆194Updated 6 months ago
- Official PyTorch Code for "Is Synthetic Data From Diffusion Models Ready for Knowledge Distillation?" (https://arxiv.org/abs/2305.12954)☆47Updated last year
- AlphaEdit: Null-Space Constrained Knowledge Editing for Language Models, ICLR 2025 (Outstanding Paper)☆214Updated 3 weeks ago
- QuEST: Efficient Finetuning for Low-bit Diffusion Models☆44Updated 3 months ago
- [CVPR'23] SparseViT: Revisiting Activation Sparsity for Efficient High-Resolution Vision Transformer☆68Updated last year
- [ICML'24 Oral] APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and Inference☆39Updated 11 months ago
- Official code implementation for 2025 ICLR accepted paper "Dobi-SVD : Differentiable SVD for LLM Compression and Some New Perspectives"☆31Updated last month
- ☆9Updated last year
- Awesome list for LLM pruning.☆224Updated 5 months ago
- A generalized framework for subspace tuning methods in parameter efficient fine-tuning.☆141Updated 3 months ago
- A library for calculating the FLOPs in the forward() process based on torch.fx☆109Updated last month
- Pytorch implementation of our paper accepted by CVPR 2022 -- IntraQ: Learning Synthetic Images with Intra-Class Heterogeneity for Zero-Sh…☆32Updated 3 years ago
- ☆11Updated 5 months ago
- Prioritize Alignment in Dataset Distillation☆20Updated 5 months ago