withinmiaov / A-Survey-on-Mixture-of-Experts
☆146Updated last month
Related projects ⓘ
Alternatives and complementary repositories for A-Survey-on-Mixture-of-Experts
- ☆116Updated 3 months ago
- ☆131Updated 2 months ago
- A curated list of Model Merging methods.☆83Updated 2 months ago
- ☆76Updated 4 months ago
- [ICLR 2024 Spotlight] Code for the paper "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy"☆64Updated 5 months ago
- Survey on Data-centric Large Language Models☆65Updated 4 months ago
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆74Updated this week
- ☆58Updated 3 months ago
- Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. arXiv:2408.07666.☆217Updated this week
- [EMNLP 2023 Main] Sparse Low-rank Adaptation of Pre-trained Language Models☆69Updated 8 months ago
- Continual Learning of Large Language Models: A Comprehensive Survey☆252Updated last week
- [NeurIPS 24 Spotlight] MaskLLM: Learnable Semi-structured Sparsity for Large Language Models☆118Updated 3 weeks ago
- Code accompanying the paper "Massive Activations in Large Language Models"☆123Updated 8 months ago
- LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment☆230Updated 6 months ago
- Survey Paper List - Efficient LLM and Foundation Models☆220Updated last month
- Awesome-Low-Rank-Adaptation☆38Updated last month
- ☆147Updated 4 months ago
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆92Updated 2 months ago
- Official implementation of "DoRA: Weight-Decomposed Low-Rank Adaptation"☆123Updated 6 months ago
- [NeurIPS2024] Twin-Merging: Dynamic Integration of Modular Expertise in Model Merging☆36Updated this week
- A curated list of awesome Multimodal studies.☆97Updated 2 weeks ago
- A generalized framework for subspace tuning methods in parameter efficient fine-tuning.☆102Updated 2 months ago
- ☆154Updated last month
- [EMNLP 2024 Findings🔥] Official implementation of "LOOK-M: Look-Once Optimization in KV Cache for Efficient Multimodal Long-Context Infe…☆75Updated last week
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆74Updated 3 weeks ago
- Collection of Tools and Papers related to Adapters / Parameter-Efficient Transfer Learning/ Fine-Tuning☆176Updated 6 months ago
- Awesome list for LLM pruning.☆167Updated this week
- The official implementation of the paper <MoA: Mixture of Sparse Attention for Automatic Large Language Model Compression>☆100Updated last week
- [NeurIPS 2024 Spotlight] EMR-Merging: Tuning-Free High-Performance Model Merging☆31Updated 3 weeks ago
- [ICML 2024 Oral] This project is the official implementation of our Accurate LoRA-Finetuning Quantization of LLMs via Information Retenti…☆59Updated 7 months ago