shawnricecake / HeimaLinks
Code for Heima
☆43Updated last month
Alternatives and similar repositories for Heima
Users that are interested in Heima are comparing it to the libraries listed below
Sorting:
- [ICML 2025] M-STAR (Multimodal Self-Evolving TrAining for Reasoning) Project. Diving into Self-Evolving Training for Multimodal Reasoning☆60Updated 5 months ago
- Github repository for "Bring Reason to Vision: Understanding Perception and Reasoning through Model Merging" (ICML 2025)☆51Updated last week
- NoisyRollout: Reinforcing Visual Reasoning with Data Augmentation☆64Updated 2 weeks ago
- Code for paper "Unraveling Cross-Modality Knowledge Conflicts in Large Vision-Language Models."☆42Updated 7 months ago
- Enhancing Large Vision Language Models with Self-Training on Image Comprehension.☆67Updated last year
- Think or Not? Selective Reasoning via Reinforcement Learning for Vision-Language Models☆36Updated 2 weeks ago
- ☆27Updated last month
- A Sober Look at Language Model Reasoning☆52Updated last week
- ☆29Updated last year
- Codes for Merging Large Language Models☆31Updated 9 months ago
- SIFT: Grounding LLM Reasoning in Contexts via Stickers☆56Updated 3 months ago
- Optimizing Anytime Reasoning via Budget Relative Policy Optimization☆36Updated last week
- Unofficial Implementation of Chain-of-Thought Reasoning Without Prompting☆32Updated last year
- Evaluation framework for paper "VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?"☆57Updated 7 months ago
- Official repository for Montessori-Instruct: Generate Influential Training Data Tailored for Student Learning [ICLR 2025]☆45Updated 4 months ago
- X-Reasoner: Towards Generalizable Reasoning Across Modalities and Domains☆43Updated 3 weeks ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆69Updated 3 months ago
- Official implementation of the paper "MMInA: Benchmarking Multihop Multimodal Internet Agents"☆43Updated 3 months ago
- [ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibrati…☆38Updated 11 months ago
- ☆105Updated 2 months ago
- ☆17Updated 5 months ago
- [ICLR 2025] When Attention Sink Emerges in Language Models: An Empirical View (Spotlight)☆85Updated 7 months ago
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆70Updated 2 months ago
- Code for "Reasoning to Learn from Latent Thoughts"☆104Updated 2 months ago
- ☆15Updated last month
- [𝐄𝐌𝐍𝐋𝐏 𝐅𝐢𝐧𝐝𝐢𝐧𝐠𝐬 𝟐𝟎𝟐𝟒 & 𝐀𝐂𝐋 𝟐𝟎𝟐𝟒 𝐍𝐋𝐑𝐒𝐄 𝐎𝐫𝐚𝐥] 𝘌𝘯𝘩𝘢𝘯𝘤𝘪𝘯𝘨 𝘔𝘢𝘵𝘩𝘦𝘮𝘢𝘵𝘪𝘤𝘢𝘭 𝘙𝘦𝘢𝘴𝘰𝘯𝘪𝘯…☆51Updated last year
- ☆28Updated last month
- [NAACL 2025 Oral] Multimodal Needle in a Haystack (MMNeedle): Benchmarking Long-Context Capability of Multimodal Large Language Models☆44Updated last month
- What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆64Updated 3 months ago
- [ACL 2025] A Generalizable and Purely Unsupervised Self-Training Framework☆60Updated this week