jadohu / LANTERNLinks
Official Implementation of LANTERN (ICLR'25) and LANTERN++(ICLRW-SCOPE'25)
☆19Updated 11 months ago
Alternatives and similar repositories for LANTERN
Users that are interested in LANTERN are comparing it to the libraries listed below
Sorting:
- Awesome-Low-Rank-Adaptation☆128Updated last year
- A curated reading list of research in Mixture-of-Experts(MoE).☆659Updated last year
- [ICML‘24] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".☆123Updated 7 months ago
- Source code of "Task arithmetic in the tangent space: Improved editing of pre-trained models".☆110Updated 2 years ago
- ☆14Updated last year
- [ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Di…☆70Updated last year
- Compressible Dynamics in Deep Overparameterized Low-Rank Learning & Adaptation (ICML'24 Oral)☆13Updated last year
- Official PyTorch implementation of DistiLLM-2: A Contrastive Approach Boosts the Distillation of LLMs (ICML 2025 Oral)☆55Updated 7 months ago
- ☆13Updated 3 months ago
- ☆28Updated 11 months ago
- Implementations for "Trimming the ℓ₁ Regularizer: Statistical Analysis, Optimization, and Applications to Deep Learning" Published on ICM…☆34Updated 5 years ago
- [TKDE'25] The official GitHub page for the survey paper "A Survey on Mixture of Experts in Large Language Models".☆482Updated 6 months ago
- ☆56Updated last year
- Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. ACM Computing Surveys, 2025.☆659Updated last week
- LoRA-XS: Low-Rank Adaptation with Extremely Small Number of Parameters☆45Updated 6 months ago
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆233Updated last year
- 📰 Must-read papers and blogs on Speculative Decoding ⚡️☆1,117Updated 2 weeks ago
- Second-Order Fine-Tuning without Pain for LLMs: a Hessian Informed Zeroth-Order Optimizer☆23Updated 11 months ago
- ThinK: Thinner Key Cache by Query-Driven Pruning☆27Updated 11 months ago
- An implementation of the DISP-LLM method from the NeurIPS 2024 paper: Dimension-Independent Structural Pruning for Large Language Models.☆23Updated 6 months ago
- Compressed LLMs for Efficient Text Generation [ICLR'24 Workshop]☆90Updated last year
- Welcome to the 'In Context Learning Theory' Reading Group☆30Updated last year
- Official Pytorch Implementation of "Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity"☆80Updated 7 months ago
- 🔨 Malet (Machine Learning Experiment Tool) is a tool for efficient machine learning experiment execution, logging, analysis, and plot ma…☆17Updated 2 weeks ago
- ☆63Updated last year
- ☆32Updated last month
- Code accompanying the paper "Massive Activations in Large Language Models"☆195Updated last year
- A curated list of early exiting (LLM, CV, NLP, etc)☆70Updated last year
- ☆17Updated 3 years ago
- Dataset Condensation (ICLR21 and ICML21)☆544Updated 2 years ago