THUDM / MoELoRA_RiemannianLinks
Source code of paper: A Stronger Mixture of Low-Rank Experts for Fine-Tuning Foundation Models. (ICML 2025)
☆35Updated 9 months ago
Alternatives and similar repositories for MoELoRA_Riemannian
Users that are interested in MoELoRA_Riemannian are comparing it to the libraries listed below
Sorting:
- [NeurIPS 2025] Think or Not? Selective Reasoning via Reinforcement Learning for Vision-Language Models☆52Updated 3 months ago
- [NeurIPS 2024] A Novel Rank-Based Metric for Evaluating Large Language Models☆56Updated 7 months ago
- [EMNLP 2025 Main] AlphaOne: Reasoning Models Thinking Slow and Fast at Test Time☆89Updated 7 months ago
- ☆36Updated 3 months ago
- ☆53Updated 10 months ago
- ☆41Updated 4 months ago
- ☆46Updated 4 months ago
- Github repository for "Why Is Spatial Reasoning Hard for VLMs? An Attention Mechanism Perspective on Focus Areas" (ICML 2025)☆67Updated 8 months ago
- R1-Searcher++: Incentivizing the Dynamic Knowledge Acquisition of LLMs via Reinforcement Learning☆69Updated 7 months ago
- ☆46Updated 2 months ago
- ☆23Updated 7 months ago
- ☆176Updated last month
- Offical Repository of "AtomThink: Multimodal Slow Thinking with Atomic Step Reasoning"☆58Updated last month
- This repository will continuously update the latest papers, technical reports, benchmarks about multimodal reasoning!☆53Updated 9 months ago
- [NeurIPS 2025] Unsupervised Post-Training for Multi-Modal LLM Reasoning via GRPO☆73Updated 2 months ago
- ☆109Updated last year
- [EMNLP 2024 Findings] Unlocking Continual Learning Abilities in Language Models☆26Updated last year
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆109Updated 7 months ago
- The code and data of We-Math, accepted by ACL 2025 main conference.☆134Updated last month
- JudgeLRM: Large Reasoning Models as a Judge☆40Updated last month
- Official repo for "Spotlight on Token Perception for Multimodal Reinforcement Learning"☆45Updated 2 months ago
- Test-time preferenece optimization (ICML 2025).☆177Updated 8 months ago
- AdaMoLE: Adaptive Mixture of LoRA Experts☆38Updated last year
- [ACL'25] We propose a novel fine-tuning method, Separate Memory and Reasoning, which combines prompt tuning with LoRA.☆82Updated 2 months ago
- ☆51Updated 8 months ago
- CPPO: Accelerating the Training of Group Relative Policy Optimization-Based Reasoning Models (NeurIPS 2025)☆171Updated 2 months ago
- ☆32Updated 5 months ago
- An benchmark for evaluating the capabilities of large vision-language models (LVLMs)☆46Updated 2 years ago
- RewardAnything: Generalizable Principle-Following Reward Models☆45Updated 7 months ago
- Parameter-Efficient Fine-Tuning for Foundation Models☆105Updated 9 months ago