THUDM / MoELoRA_RiemannianLinks
Source code of paper: A Stronger Mixture of Low-Rank Experts for Fine-Tuning Foundation Models. (ICML 2025)
☆32Updated 4 months ago
Alternatives and similar repositories for MoELoRA_Riemannian
Users that are interested in MoELoRA_Riemannian are comparing it to the libraries listed below
Sorting:
- Think or Not? Selective Reasoning via Reinforcement Learning for Vision-Language Models☆40Updated 2 weeks ago
- [EMNLP 2025 Main] AlphaOne: Reasoning Models Thinking Slow and Fast at Test Time☆77Updated 2 months ago
- This repository will continuously update the latest papers, technical reports, benchmarks about multimodal reasoning!☆48Updated 5 months ago
- ☆86Updated 7 months ago
- RewardAnything: Generalizable Principle-Following Reward Models☆34Updated 2 months ago
- ☆120Updated 5 months ago
- ☆147Updated 3 months ago
- Code for paper "Unraveling Cross-Modality Knowledge Conflicts in Large Vision-Language Models."☆45Updated 10 months ago
- codes for Efficient Test-Time Scaling via Self-Calibration☆18Updated 5 months ago
- ☆48Updated 3 months ago
- AdaMoLE: Adaptive Mixture of LoRA Experts☆36Updated 10 months ago
- ☆96Updated last month
- Extensive Self-Contrast Enables Feedback-Free Language Model Alignment☆20Updated last year
- A Self-Training Framework for Vision-Language Reasoning☆82Updated 7 months ago
- A curated list of awesome LLM Inference-Time Self-Improvement (ITSI, pronounced "itsy") papers from our recent survey: A Survey on Large …☆95Updated 8 months ago
- [ICLR 2025] Mitigating Modality Prior-Induced Hallucinations in Multimodal Large Language Models via Deciphering Attention Causality☆37Updated last month
- An benchmark for evaluating the capabilities of large vision-language models (LVLMs)☆46Updated last year
- Test-time preferenece optimization (ICML 2025).☆162Updated 3 months ago
- ☆19Updated 3 months ago
- SFT or RL? An Early Investigation into Training R1-Like Reasoning Large Vision-Language Models☆132Updated 4 months ago
- ☆80Updated last year
- ☆50Updated 5 months ago
- Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimization☆91Updated last year
- Offical Repository of "AtomThink: Multimodal Slow Thinking with Atomic Step Reasoning"☆54Updated 3 weeks ago
- Official implementation of "Automated Generation of Challenging Multiple-Choice Questions for Vision Language Model Evaluation" (CVPR 202…☆33Updated 3 months ago
- Reinforced Multi-LLM Agents training☆39Updated 2 months ago
- ☆21Updated 9 months ago
- ☆113Updated last year
- DeepPerception: Advancing R1-like Cognitive Visual Perception in MLLMs for Knowledge-Intensive Visual Grounding☆65Updated 2 months ago
- An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT☆111Updated 5 months ago