An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT
☆139Mar 11, 2025Updated last year
Alternatives and similar repositories for MoE-PEFT
Users that are interested in MoE-PEFT are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆204Aug 22, 2024Updated last year
- An Efficient "Factory" to Build Multiple LoRA Adapters☆376Feb 13, 2025Updated last year
- This repository has transferred to https://github.com/TUDB-Labs/MoE-PEFT☆22Aug 16, 2024Updated last year
- [SIGIR'24] The official implementation code of MOELoRA.☆192Jul 22, 2024Updated last year
- LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment☆404Apr 29, 2024Updated 2 years ago
- Deploy to Railway using AI coding agents - Free Credits Offer • AdUse Claude Code, Codex, OpenCode, and more. Autonomous software development now has the infrastructure to match with Railway.
- ☆179Jul 22, 2024Updated last year
- ☆68Dec 2, 2024Updated last year
- X-LoRA: Mixture of LoRA Experts☆270Aug 4, 2024Updated last year
- ☆126Jul 6, 2024Updated last year
- A list of papers for group meeting☆19Mar 6, 2026Updated 2 months ago
- Adapt an LLM model to a Mixture-of-Experts model using Parameter Efficient finetuning (LoRA), injecting the LoRAs in the FFN.☆84Oct 21, 2025Updated 6 months ago
- Xmixers: A collection of SOTA efficient token/channel mixers☆28Sep 4, 2025Updated 8 months ago
- [ ICLR 2025 ] Making LLMs More Effective with Hierarchical Mixture of LoRA Experts☆31Oct 9, 2025Updated 7 months ago
- Learnable Global Pooling Layers Based on Regularized Optimal Transport (ROT)☆16Mar 17, 2024Updated 2 years ago
- Managed Kubernetes at scale on DigitalOcean • AdDigitalOcean Kubernetes includes the control plane, bandwidth allowance, container registry, automatic updates, and more for free.
- Grounding Language Models for Compositional and Spatial Reasoning☆18Oct 26, 2022Updated 3 years ago
- ☆25Nov 25, 2025Updated 5 months ago
- [AAAI 2023] IterDE: An Iterative Knowledge Distillation Framework for Knowledge Graph Embeddings☆10Apr 3, 2024Updated 2 years ago
- [ICCV 2025 Highlight] Official code for UnZipLoRA: Separating Content and Style from a Single Image☆40Jul 30, 2025Updated 9 months ago
- ACL 2024 (SRW), Official Codebase of our Paper: "MoExtend: Tuning New Experts for Modality and Task Extension"☆15Dec 3, 2024Updated last year
- ☆64Oct 17, 2023Updated 2 years ago
- ☆16Mar 20, 2026Updated last month
- ISP^2 is a plug-and-play prompting method☆12Jun 24, 2025Updated 10 months ago
- [ICLR 2024] Beyond Accuracy: Evaluating Self-Consistency of Code Large Language Models with IdentityChain☆10Nov 24, 2025Updated 5 months ago
- Serverless GPU API endpoints on Runpod - Get Bonus Credits • AdSkip the infrastructure headaches. Auto-scaling, pay-as-you-go, no-ops approach lets you focus on innovating your application.
- A library for easily merging multiple LLM experts, and efficiently train the merged LLM.☆512Aug 26, 2024Updated last year
- PiSSA: Principal Singular Values and Singular Vectors Adaptation of Large Language Models(NeurIPS 2024 Spotlight)☆424Jun 30, 2025Updated 10 months ago
- Wiener Graph Deconvolutional Network Improves Self-Supervised Learning in AAAI 2023☆18Apr 3, 2024Updated 2 years ago
- ☆18Aug 11, 2022Updated 3 years ago
- ☆14May 9, 2024Updated 2 years ago
- ☆26Nov 23, 2023Updated 2 years ago
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆148Apr 8, 2025Updated last year
- flow-merge is a powerful Python library that enables seamless merging of multiple transformer-based language models using the most popula…☆20Feb 12, 2025Updated last year
- Mixture of Decision Trees for Interpretable Machine Learning☆11Sep 2, 2021Updated 4 years ago
- Wordpress hosting with auto-scaling - Free Trial Offer • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆120Apr 13, 2026Updated 3 weeks ago
- This repository contains the implementation of the paper "MeteoRA: Multiple-tasks Embedded LoRA for Large Language Models".☆25May 28, 2025Updated 11 months ago
- ☆153Sep 9, 2024Updated last year
- ☆54May 8, 2023Updated 3 years ago
- The predecessor of CiteLab.☆18Feb 3, 2026Updated 3 months ago
- [LREC-COLING'24] HumanEval-XL: A Multilingual Code Generation Benchmark for Cross-lingual Natural Language Generalization☆41Mar 7, 2025Updated last year
- This is a unified platform for implementing and evaluating test-time reasoning mechanisms in Large Language Models (LLMs).☆19Jan 16, 2025Updated last year