THUDM / MoELoRA_RiemannianLinks
Source code of paper: A Stronger Mixture of Low-Rank Experts for Fine-Tuning Foundation Models. (ICML 2025)
☆33Updated 6 months ago
Alternatives and similar repositories for MoELoRA_Riemannian
Users that are interested in MoELoRA_Riemannian are comparing it to the libraries listed below
Sorting:
- ☆50Updated 2 months ago
- [EMNLP 2024 Findings] Unlocking Continual Learning Abilities in Language Models☆25Updated last year
- [EMNLP 2025 Main] AlphaOne: Reasoning Models Thinking Slow and Fast at Test Time☆84Updated 4 months ago
- AdaMoLE: Adaptive Mixture of LoRA Experts☆37Updated last year
- ☆38Updated 2 months ago
- [NeurIPS 2025] Think or Not? Selective Reasoning via Reinforcement Learning for Vision-Language Models☆47Updated 2 weeks ago
- ☆50Updated 7 months ago
- [arxiv: 2505.02156] Adaptive Thinking via Mode Policy Optimization for Social Language Agents☆45Updated 3 months ago
- ☆40Updated 2 months ago
- [NeurIPS 2024] A Novel Rank-Based Metric for Evaluating Large Language Models☆54Updated 4 months ago
- codes for Efficient Test-Time Scaling via Self-Calibration☆19Updated last month
- ☆22Updated 5 months ago
- A curated list of awesome LLM Inference-Time Self-Improvement (ITSI, pronounced "itsy") papers from our recent survey: A Survey on Large …☆96Updated 9 months ago
- ☆36Updated last week
- ☆22Updated last year
- This repository will continuously update the latest papers, technical reports, benchmarks about multimodal reasoning!☆53Updated 6 months ago
- A Self-Training Framework for Vision-Language Reasoning☆86Updated 8 months ago
- ☆94Updated 9 months ago
- ☆156Updated last week
- [ACL 2025] Knowledge Unlearning for Large Language Models☆43Updated last month
- Code for paper "Unraveling Cross-Modality Knowledge Conflicts in Large Vision-Language Models."☆46Updated last year
- ☆21Updated 11 months ago
- [ACL'25] We propose a novel fine-tuning method, Separate Memory and Reasoning, which combines prompt tuning with LoRA.☆76Updated last month
- Adapt an LLM model to a Mixture-of-Experts model using Parameter Efficient finetuning (LoRA), injecting the LoRAs in the FFN.☆61Updated last year
- ☆116Updated last year
- ☆59Updated last month
- The official repository of "SmartAgent: Chain-of-User-Thought for Embodied Personalized Agent in Cyber World".☆27Updated last month
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆131Updated 11 months ago
- [NeurIPS 2025] Think Silently, Think Fast: Dynamic Latent Compression of LLM Reasoning Chains☆57Updated 2 months ago
- ICLR 2025☆29Updated 4 months ago