EricLBuehler / xlora
X-LoRA: Mixture of LoRA Experts
☆201Updated 5 months ago
Alternatives and similar repositories for xlora:
Users that are interested in xlora are comparing it to the libraries listed below
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆139Updated 4 months ago
- ☆250Updated last year
- Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.☆387Updated 9 months ago
- ☆160Updated 11 months ago
- This is the official repository for Inheritune.☆109Updated 3 months ago
- Code for In-context Vectors: Making In Context Learning More Effective and Controllable Through Latent Space Steering☆156Updated 3 months ago
- Benchmarking LLMs with Challenging Tasks from Real Users☆208Updated 2 months ago
- ☆180Updated this week
- ☆125Updated last year
- ☆214Updated 7 months ago
- Code and Data for "Long-context LLMs Struggle with Long In-context Learning"☆97Updated 6 months ago
- ☆56Updated 2 weeks ago
- ☆192Updated last month
- The official implementation of the paper "What Matters in Transformers? Not All Attention is Needed".☆158Updated last month
- The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction☆378Updated 6 months ago
- Official repository of NEFTune: Noisy Embeddings Improves Instruction Finetuning☆388Updated 8 months ago
- ☆142Updated last week
- Implementation of DoRA☆288Updated 7 months ago
- Unofficial implementation for the paper "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆145Updated 7 months ago
- Implementation of 🥥 Coconut, Chain of Continuous Thought, in Pytorch☆150Updated 3 weeks ago
- Multipack distributed sampler for fast padding-free training of LLMs☆184Updated 5 months ago
- DSIR large-scale data selection framework for language model training☆242Updated 9 months ago
- ☆209Updated 7 months ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning☆347Updated 4 months ago
- Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmind☆173Updated 4 months ago
- LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.☆123Updated 5 months ago
- MoRA: High-Rank Updating for Parameter-Efficient Fine-Tuning☆349Updated 5 months ago
- [ICLR 2024 Spotlight] Code for the paper "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy"☆70Updated 7 months ago
- Implementation of paper Data Engineering for Scaling Language Models to 128K Context☆450Updated 10 months ago
- "Improving Mathematical Reasoning with Process Supervision" by OPENAI☆103Updated this week