MoRA: High-Rank Updating for Parameter-Efficient Fine-Tuning
☆361Aug 7, 2024Updated last year
Alternatives and similar repositories for MoRA
Users that are interested in MoRA are comparing it to the libraries listed below
Sorting:
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆203Jul 17, 2024Updated last year
- Official code for ReLoRA from the paper Stack More Layers Differently: High-Rank Training Through Low-Rank Updates☆473Apr 21, 2024Updated last year
- PiSSA: Principal Singular Values and Singular Vectors Adaptation of Large Language Models(NeurIPS 2024 Spotlight)☆409Jun 30, 2025Updated 8 months ago
- [ICML2024 (Oral)] Official PyTorch implementation of DoRA: Weight-Decomposed Low-Rank Adaptation☆940Oct 1, 2024Updated last year
- Stanford NLP Python library for Representation Finetuning (ReFT)☆1,560Jan 14, 2026Updated last month
- ☆233Jun 24, 2024Updated last year
- The code repository for the CURLoRA research paper. Stable LLM continual fine-tuning and catastrophic forgetting mitigation.☆53Aug 28, 2024Updated last year
- KURE: 고려대학교에서 개발한, 한국어 검색에 특화된 임베딩 모델☆206Feb 20, 2026Updated last week
- GRadient-INformed MoE☆264Sep 25, 2024Updated last year
- Block Transformer: Global-to-Local Language Modeling for Fast Inference (NeurIPS 2024)☆163Apr 13, 2025Updated 10 months ago
- Code for paper "Patch-Level Training for Large Language Models"☆96Nov 10, 2025Updated 3 months ago
- Implementation of DoRA☆307Jun 7, 2024Updated last year
- Modified Beam Search with periodical restart☆12Sep 12, 2024Updated last year
- Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793☆453May 13, 2025Updated 9 months ago
- [KO-Platy🥮] Korean-Open-platypus를 활용하여 llama-2-ko를 fine-tuning한 KO-platypus model☆73Aug 24, 2025Updated 6 months ago
- LLM 모델의 외국어 토큰 생성을 막는 코드 구현☆83Aug 7, 2025Updated 6 months ago
- 언어모델을 학습하기 위한 공개 한국어 instruction dataset들을 모아두었습니다.☆453Apr 13, 2025Updated 10 months ago
- Layer-Condensed KV cache w/ 10 times larger batch size, fewer params and less computation. Dramatic speed up with better task performance…☆156Apr 7, 2025Updated 10 months ago
- ☆54Sep 27, 2024Updated last year
- A library for easily merging multiple LLM experts, and efficiently train the merged LLM.☆507Aug 26, 2024Updated last year
- Tools for merging pretrained large language models.☆6,826Updated this week
- DeMo: Decoupled Momentum Optimization☆198Dec 2, 2024Updated last year
- The official implementation of Self-Play Fine-Tuning (SPIN)☆1,235May 8, 2024Updated last year
- The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction☆390Jul 9, 2024Updated last year
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆144Apr 8, 2025Updated 10 months ago
- GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection☆1,678Oct 28, 2024Updated last year
- ☆235Jun 11, 2024Updated last year
- [ICML 2024] CLLMs: Consistency Large Language Models☆412Nov 16, 2024Updated last year
- ☆130Oct 1, 2024Updated last year
- XmodelLM☆38Nov 19, 2024Updated last year
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆946Feb 16, 2025Updated last year
- [TMLR 2025] When Attention Collapses: How Degenerate Layers in LLMs Enable Smaller, Stronger Models☆125Feb 15, 2026Updated 2 weeks ago
- ☆217Nov 25, 2025Updated 3 months ago
- ☆15Mar 12, 2024Updated last year
- [ICLR 2024] This is the repository for the paper titled "DePT: Decomposed Prompt Tuning for Parameter-Efficient Fine-tuning"☆101Apr 10, 2024Updated last year
- Some preliminary explorations of Mamba's context scaling.☆218Feb 8, 2024Updated 2 years ago
- ☆301Jul 10, 2025Updated 7 months ago
- 한국어 언어모델 다분야 사고력 벤치마크☆201Oct 17, 2024Updated last year
- A high-throughput and memory-efficient inference and serving engine for LLMs☆267Dec 4, 2025Updated 2 months ago