yushuiwx / Mixture-of-LoRA-Experts
☆33Updated 5 months ago
Alternatives and similar repositories for Mixture-of-LoRA-Experts
Users that are interested in Mixture-of-LoRA-Experts are comparing it to the libraries listed below
Sorting:
- ☆134Updated 9 months ago
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆199Updated 5 months ago
- MoCLE (First MLLM with MoE for instruction customization and generalization!) (https://arxiv.org/abs/2312.12379)☆37Updated last year
- ☆101Updated 10 months ago
- TRACE: A Comprehensive Benchmark for Continual Learning in Large Language Models☆69Updated last year
- ☆174Updated 10 months ago
- [EMNLP 2024] mDPO: Conditional Preference Optimization for Multimodal Large Language Models.☆73Updated 6 months ago
- [SIGIR'24] The official implementation code of MOELoRA.☆162Updated 9 months ago
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆141Updated 2 months ago
- Code for ACL 2024 accepted paper titled "SAPT: A Shared Attention Framework for Parameter-Efficient Continual Learning of Large Language …☆34Updated 4 months ago
- [NeurIPS 2023] Github repository for "Composing Parameter-Efficient Modules with Arithmetic Operations"☆61Updated last year
- ☆194Updated 6 months ago
- ☆97Updated 2 months ago
- [ICML'2024] Can AI Assistants Know What They Don't Know?☆81Updated last year
- ☆29Updated last year
- [ICLR 2025] Released code for paper "Spurious Forgetting in Continual Learning of Language Models"☆40Updated last week
- ☆73Updated 11 months ago
- [EMNLP 2024 Findings🔥] Official implementation of ": LOOK-M: Look-Once Optimization in KV Cache for Efficient Multimodal Long-Context In…☆93Updated 6 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆67Updated 3 months ago
- AdaMoLE: Adaptive Mixture of LoRA Experts☆29Updated 7 months ago
- Code for paper "UniPELT: A Unified Framework for Parameter-Efficient Language Model Tuning", ACL 2022☆59Updated 3 years ago
- my commonly-used tools☆53Updated 4 months ago
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆119Updated 6 months ago
- An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT☆97Updated 2 months ago
- [ICLR 2025] 🧬 RegMix: Data Mixture as Regression for Language Model Pre-training (Spotlight)☆133Updated 2 months ago
- Model merging is a highly efficient approach for long-to-short reasoning.☆46Updated last month
- LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment☆331Updated last year
- ☆18Updated 5 months ago
- ☆43Updated last month
- ☆26Updated 4 months ago