TUDB-Labs / MixLoRA
State-of-the-art Parameter-Efficient MoE Fine-tuning Method
☆161Updated 8 months ago
Alternatives and similar repositories for MixLoRA
Users that are interested in MixLoRA are comparing it to the libraries listed below
Sorting:
- An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT☆97Updated 2 months ago
- ☆134Updated 9 months ago
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆141Updated 2 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆67Updated 3 months ago
- ☆101Updated 10 months ago
- [SIGIR'24] The official implementation code of MOELoRA.☆162Updated 9 months ago
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆89Updated 3 months ago
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆199Updated 5 months ago
- ☆97Updated 2 months ago
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆119Updated 6 months ago
- LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment☆332Updated last year
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆202Updated this week
- ☆79Updated 3 weeks ago
- ☆145Updated 8 months ago
- ☆194Updated 6 months ago
- AlphaEdit: Null-Space Constrained Knowledge Editing for Language Models, ICLR 2025 (Outstanding Paper)☆214Updated 3 weeks ago
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆116Updated last month
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆119Updated last month
- ☆174Updated 10 months ago
- Official repository for paper: O1-Pruner: Length-Harmonizing Fine-Tuning for O1-Like Reasoning Pruning☆74Updated 2 months ago
- [Arxiv 2025] Efficient Reasoning Models: A Survey☆146Updated last week
- [ICLR 2025] Code and Data Repo for Paper "Latent Space Chain-of-Embedding Enables Output-free LLM Self-Evaluation"☆46Updated 4 months ago
- 🚀 LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training☆83Updated 5 months ago
- This repository contains a regularly updated paper list for LLMs-reasoning-in-latent-space.☆80Updated last week
- AdaMerging: Adaptive Model Merging for Multi-Task Learning. ICLR, 2024.☆80Updated 6 months ago
- ☆82Updated last week
- ☆106Updated last week
- ☆24Updated 2 months ago
- Model merging is a highly efficient approach for long-to-short reasoning.☆46Updated last month
- ☆24Updated 11 months ago