NJUDeepEngine / meteoraLinks
This repository contains the implementation of the paper "MeteoRA: Multiple-tasks Embedded LoRA for Large Language Models".
β20Updated last month
Alternatives and similar repositories for meteora
Users that are interested in meteora are comparing it to the libraries listed below
Sorting:
- Preprint: Asymmetry in Low-Rank Adapters of Foundation Modelsβ35Updated last year
- [ICLR 2025π₯] D2O: Dynamic Discriminative Operations for Efficient Long-Context Inference of Large Language Modelsβ16Updated 3 months ago
- A Sober Look at Language Model Reasoningβ74Updated last week
- Less is More: Task-aware Layer-wise Distillation for Language Model Compression (ICML2023)β35Updated last year
- AdaMerging: Adaptive Model Merging for Multi-Task Learning. ICLR, 2024.β83Updated 8 months ago
- [ICML 2024] SPP: Sparsity-Preserved Parameter-Efficient Fine-Tuning for Large Language Modelsβ21Updated last year
- β57Updated last year
- [ICLR 2025] When Attention Sink Emerges in Language Models: An Empirical View (Spotlight)β88Updated 8 months ago
- [ICML 2024] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".β105Updated 11 months ago
- A family of efficient edge language models in 100M~1B sizes.β15Updated 4 months ago
- Test-time-training on nearest neighbors for large language modelsβ41Updated last year
- PoC for "SpecReason: Fast and Accurate Inference-Time Compute via Speculative Reasoning" [arXiv '25]β39Updated last month
- β18Updated 7 months ago
- β52Updated 6 months ago
- Representation Surgery for Multi-Task Model Merging. ICML, 2024.β45Updated 8 months ago
- β71Updated 7 months ago
- β49Updated last year
- β104Updated 3 weeks ago
- [ICLR 2024 Spotlight] Code for the paper "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy"β85Updated last week
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Modelsβ110Updated 4 months ago
- [ICML 2024] Junk DNA Hypothesis: A Task-Centric Angle of LLM Pre-trained Weights through Sparsity; Lu Yin*, Ajay Jaiswal*, Shiwei Liu, Soβ¦β16Updated 2 months ago
- β30Updated last month
- Code for "ECoFLaP: Efficient Coarse-to-Fine Layer-Wise Pruning for Vision-Language Models" (ICLR 2024)β19Updated last year
- Implementation of CoLA: Compute-Efficient Pre-Training of LLMs via Low-Rank Activationβ22Updated 4 months ago
- [ICLR 2025 Workshop] "Landscape of Thoughts: Visualizing the Reasoning Process of Large Language Models"β25Updated last week
- [ACL 2024] Not All Experts are Equal: Efficient Expert Pruning and Skipping for Mixture-of-Experts Large Language Modelsβ92Updated last year
- [NeurIPS 2024 Spotlight] EMR-Merging: Tuning-Free High-Performance Model Mergingβ59Updated 3 months ago
- [ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibratiβ¦β41Updated 11 months ago
- Official Pytorch Implementation of Our Paper Accepted at ICLR 2024-- Dynamic Sparse No Training: Training-Free Fine-tuning for Sparse LLMβ¦β47Updated last year
- Papers of Implicit Reasoning in LLMs.β13Updated 3 months ago