withinmiaov / A-Survey-on-Mixture-of-Experts-in-LLMs
The official GitHub page for the survey paper "A Survey on Mixture of Experts in Large Language Models".
☆316Updated 3 weeks ago
Alternatives and similar repositories for A-Survey-on-Mixture-of-Experts-in-LLMs:
Users that are interested in A-Survey-on-Mixture-of-Experts-in-LLMs are comparing it to the libraries listed below
- Paper list for Efficient Reasoning.☆362Updated this week
- A Survey on Efficient Reasoning for LLMs☆281Updated this week
- Awesome list for LLM pruning.☆218Updated 3 months ago
- 📰 Must-read papers on KV Cache Compression (constantly updating 🤗).☆367Updated last week
- Survey Paper List - Efficient LLM and Foundation Models☆241Updated 6 months ago
- Awesome LLM pruning papers all-in-one repository with integrating all useful resources and insights.☆79Updated 4 months ago
- Awesome list for LLM quantization☆195Updated 3 months ago
- Paper List of Inference/Test Time Scaling/Computing☆160Updated this week
- 😎 A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond☆152Updated this week
- Spec-Bench: A Comprehensive Benchmark and Unified Evaluation Platform for Speculative Decoding (ACL 2024 Findings)☆247Updated 2 weeks ago
- awesome papers in LLM interpretability☆436Updated 2 months ago
- Awesome-LLM-KV-Cache: A curated list of 📙Awesome LLM KV Cache Papers with Codes.☆266Updated last month
- A collection of 150+ surveys on LLMs☆285Updated last month
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆184Updated 4 months ago
- A collection of AWESOME things about mixture-of-experts☆1,081Updated 4 months ago
- [ICML 2024] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".☆96Updated 9 months ago
- 📰 Must-read papers and blogs on Speculative Decoding ⚡️☆672Updated last week
- [NeurIPS 24 Spotlight] MaskLLM: Learnable Semi-structured Sparsity for Large Language Models☆160Updated 3 months ago
- ☆187Updated 5 months ago
- Efficient Multimodal Large Language Models: A Survey☆331Updated last month
- LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment☆315Updated 11 months ago
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆153Updated 7 months ago
- Awesome RL-based LLM Reasoning☆406Updated this week
- A Telegram bot to recommend arXiv papers☆264Updated 2 months ago
- A curated reading list of research in Mixture-of-Experts(MoE).☆609Updated 5 months ago
- Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. arXiv:2408.07666.☆353Updated this week
- TransMLA: Multi-Head Latent Attention Is All You Need☆226Updated last month
- Survey of Small Language Models from Penn State, ...☆170Updated 2 months ago
- Continual Learning of Large Language Models: A Comprehensive Survey☆381Updated last month
- Unofficial implementation for the paper "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆156Updated 9 months ago