mst272 / simple-lora-plus
A simple implementation of LoRA+: Efficient Low Rank Adaptation of Large Models
☆9Updated last year
Alternatives and similar repositories for simple-lora-plus:
Users that are interested in simple-lora-plus are comparing it to the libraries listed below
- ☆12Updated 11 months ago
- Due to the huge vocaburary size (151,936) of Qwen models, the Embedding and LM Head weights are excessively heavy. Therefore, this projec…☆17Updated 7 months ago
- ☆106Updated 5 months ago
- Unofficial implementations of block/layer-wise pruning methods for LLMs.☆68Updated 11 months ago
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆85Updated 2 months ago
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆111Updated this week
- [ICLR 2025] MiniPLM: Knowledge Distillation for Pre-Training Language Models☆36Updated 4 months ago
- Pretrain、decay、SFT a CodeLLM from scratch 🧙♂️☆35Updated 10 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆61Updated last month
- Official Repo for SparseLLM: Global Pruning of LLMs (NeurIPS 2024)☆53Updated 2 weeks ago
- ☆131Updated 8 months ago
- code for ACL24 "MELoRA: Mini-Ensemble Low-Rank Adapter for Parameter-Efficient Fine-Tuning"☆19Updated last month
- The official PyTorch implementation of the paper "MLAE: Masked LoRA Experts for Visual Parameter-Efficient Fine-Tuning"☆27Updated 4 months ago
- A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.☆168Updated this week
- SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning. COLM 2024 Accepted Paper☆31Updated 10 months ago
- [SIGIR'24] The official implementation code of MOELoRA.☆159Updated 8 months ago
- This is a repo for showcasing using MCTS with LLMs to solve gsm8k problems☆69Updated 3 weeks ago
- ZO2 (Zeroth-Order Offloading): Full Parameter Fine-Tuning 175B LLMs with 18GB GPU Memory☆81Updated last week
- Quantized Side Tuning: Fast and Memory-Efficient Tuning of Quantized Large Language Models☆42Updated 5 months ago
- ☆165Updated this week
- ☆172Updated 9 months ago
- llm & rl☆92Updated this week
- ☆187Updated 5 months ago
- ☆112Updated this week
- A generalized framework for subspace tuning methods in parameter efficient fine-tuning.☆133Updated 2 months ago
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆117Updated 5 months ago
- LISA: Layerwise Importance Sampling for Memory-Efficient Large Language Model Fine-Tuning☆30Updated last year
- Training code for Baby-Llama, our submission to the strict-small track of the BabyLM challenge.☆79Updated last year
- D^2-MoE: Delta Decompression for MoE-based LLMs Compression☆36Updated 2 weeks ago
- The official repository for the Scientific Paper Idea Proposer (SciPIP)☆63Updated last month