Outsider565 / LoRA-GA
☆194Updated 6 months ago
Alternatives and similar repositories for LoRA-GA
Users that are interested in LoRA-GA are comparing it to the libraries listed below
Sorting:
- A generalized framework for subspace tuning methods in parameter efficient fine-tuning.☆141Updated 3 months ago
- PiSSA: Principal Singular Values and Singular Vectors Adaptation of Large Language Models(NeurIPS 2024 Spotlight)☆351Updated 3 months ago
- ☆174Updated 10 months ago
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆116Updated last month
- ☆101Updated 10 months ago
- LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment☆332Updated last year
- ☆134Updated 9 months ago
- AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).☆324Updated last year
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆199Updated 5 months ago
- ☆184Updated last month
- ☆145Updated 8 months ago
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆141Updated 2 months ago
- [SIGIR'24] The official implementation code of MOELoRA.☆162Updated 9 months ago
- qwen-nsa☆61Updated last month
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆161Updated 8 months ago
- ☆196Updated 2 months ago
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆89Updated 3 months ago
- AlphaEdit: Null-Space Constrained Knowledge Editing for Language Models, ICLR 2025 (Outstanding Paper)☆214Updated 3 weeks ago
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆119Updated 6 months ago
- ☆79Updated 3 weeks ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆67Updated 3 months ago
- An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT☆97Updated 2 months ago
- [ACL 2024] Long-Context Language Modeling with Parallel Encodings☆153Updated 11 months ago
- Rethinking RL Scaling for Vision Language Models: A Transparent, From-Scratch Framework and Comprehensive Evaluation Scheme☆122Updated last month
- Code for paper "Patch-Level Training for Large Language Models"☆84Updated 6 months ago
- [ICLR 2025] 🧬 RegMix: Data Mixture as Regression for Language Model Pre-training (Spotlight)☆135Updated 2 months ago
- ☆33Updated 5 months ago
- [EMNLP 2023, Main Conference] Sparse Low-rank Adaptation of Pre-trained Language Models☆76Updated last year
- CPPO: Accelerating the Training of Group Relative Policy Optimization-Based Reasoning Models☆125Updated last week
- [Arxiv 2025] Efficient Reasoning Models: A Survey☆146Updated last week