SuperBruceJia / Awesome-Mixture-of-ExpertsLinks
Awesome Mixture of Experts (MoE): A Curated List of Mixture of Experts (MoE) and Mixture of Multimodal Experts (MoME)
☆31Updated 6 months ago
Alternatives and similar repositories for Awesome-Mixture-of-Experts
Users that are interested in Awesome-Mixture-of-Experts are comparing it to the libraries listed below
Sorting:
- ☆113Updated 4 months ago
- [ICLR 2024] This is the repository for the paper titled "DePT: Decomposed Prompt Tuning for Parameter-Efficient Fine-tuning"☆95Updated last year
- Survey: A collection of AWESOME papers and resources on the latest research in Mixture of Experts.☆125Updated 10 months ago
- ☆108Updated last year
- ☆136Updated last month
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆116Updated last week
- A curated list of Model Merging methods.☆92Updated 10 months ago
- Parameter-Efficient Fine-Tuning for Foundation Models☆75Updated 3 months ago
- AdaMoLE: Adaptive Mixture of LoRA Experts☆33Updated 9 months ago
- [EMNLP 2023, Main Conference] Sparse Low-rank Adaptation of Pre-trained Language Models☆79Updated last year
- Must-read Papers on Large Language Model (LLM) Continual Learning☆144Updated last year
- ☆145Updated 11 months ago
- MoCLE (First MLLM with MoE for instruction customization and generalization!) (https://arxiv.org/abs/2312.12379)☆41Updated 2 weeks ago
- Collection of Tools and Papers related to Adapters / Parameter-Efficient Transfer Learning/ Fine-Tuning☆194Updated last year
- [NeurIPS 2024] A Novel Rank-Based Metric for Evaluating Large Language Models☆49Updated last month
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆126Updated 3 months ago
- 🚀 LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training☆86Updated 7 months ago
- [NeurIPS 2024 Spotlight] EMR-Merging: Tuning-Free High-Performance Model Merging☆59Updated 4 months ago
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆169Updated 10 months ago
- Awesome Low-Rank Adaptation☆39Updated last month
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆216Updated 7 months ago
- [ICCV 2025] Auto Interpretation Pipeline and many other functionalities for Multimodal SAE Analysis.☆144Updated last week
- MLLM-Bench: Evaluating Multimodal LLMs with Per-sample Criteria☆70Updated 9 months ago
- Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆99Updated last week
- [ICLR 2025] MLLM can see? Dynamic Correction Decoding for Hallucination Mitigation☆89Updated 7 months ago
- [NAACL 2025] A Closer Look into Mixture-of-Experts in Large Language Models☆52Updated 5 months ago
- [EMNLP 2024] mDPO: Conditional Preference Optimization for Multimodal Large Language Models.☆76Updated 8 months ago
- The this is the official implementation of "DAPE: Data-Adaptive Positional Encoding for Length Extrapolation"☆38Updated 9 months ago
- This repository contains the code for SFT, RLHF, and DPO, designed for vision-based LLMs, including the LLaVA models and the LLaMA-3.2-vi…☆110Updated last month
- [arXiv 2025] Efficient Reasoning Models: A Survey☆227Updated this week