ustc-hyin / HiMAPLinks
Code for paper: Unraveling the Shift of Visual Information Flow in MLLMs: From Phased Interaction to Efficient Inference
β13Updated 4 months ago
Alternatives and similar repositories for HiMAP
Users that are interested in HiMAP are comparing it to the libraries listed below
Sorting:
- [NeurIPS 2025] NoisyRollout: Reinforcing Visual Reasoning with Data Augmentationβ93Updated last month
- [EMNLP 2024 Findingsπ₯] Official implementation of ": LOOK-M: Look-Once Optimization in KV Cache for Efficient Multimodal Long-Context Inβ¦β103Updated 11 months ago
- [ICLR 2025] The official pytorch implement of "Dynamic-LLaVA: Efficient Multimodal Large Language Models via Dynamic Vision-language Contβ¦β58Updated last month
- Imagine While Reasoning in Space: Multimodal Visualization-of-Thought (ICML 2025)β53Updated 6 months ago
- [EMNLP 2024] mDPO: Conditional Preference Optimization for Multimodal Large Language Models.β82Updated 11 months ago
- This repo contains evaluation code for the paper "MileBench: Benchmarking MLLMs in Long Context"β35Updated last year
- Code release for VTW (AAAI 2025 Oral)β50Updated 3 months ago
- The official code of "VL-Rethinker: Incentivizing Self-Reflection of Vision-Language Models with Reinforcement Learning" [NeurIPS25]β157Updated 4 months ago
- [ECCV 2024] Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMsβ147Updated 11 months ago
- β109Updated last month
- Github repository for "Bring Reason to Vision: Understanding Perception and Reasoning through Model Merging" (ICML 2025)β77Updated last month
- β84Updated last year
- A Self-Training Framework for Vision-Language Reasoningβ84Updated 9 months ago
- [NeurIPS 2025] Think Silently, Think Fast: Dynamic Latent Compression of LLM Reasoning Chainsβ57Updated 2 months ago
- [ICLR 2025] MLLM can see? Dynamic Correction Decoding for Hallucination Mitigationβ118Updated last month
- Less is More: Mitigating Multimodal Hallucination from an EOS Decision Perspective (ACL 2024)β54Updated 11 months ago
- Official resource for paper Investigating and Mitigating the Multimodal Hallucination Snowballing in Large Vision-Language Models (ACL 20β¦β12Updated last year
- β58Updated 5 months ago
- LLaVA-PruMerge: Adaptive Token Reduction for Efficient Large Multimodal Modelsβ150Updated last month
- An LLM-free Multi-dimensional Benchmark for Multi-modal Hallucination Evaluationβ140Updated last year
- A RLHF Infrastructure for Vision-Language Modelsβ184Updated 11 months ago
- β¨β¨The Curse of Multi-Modalities (CMM): Evaluating Hallucinations of Large Multimodal Models across Language, Visual, and Audioβ49Updated 3 months ago
- β14Updated last year
- A paper list about Token Merge, Reduce, Resample, Drop for MLLMs.β74Updated 9 months ago
- β114Updated last week
- [ACM MM 2025] TimeChat-online: 80% Visual Tokens are Naturally Redundant in Streaming Videosβ86Updated last month
- [CVPR'2025] VoCo-LLaMA: This repo is the official implementation of "VoCo-LLaMA: Towards Vision Compression with Large Language Models".β191Updated 4 months ago
- π Global Compression Commander: Plug-and-Play Inference Acceleration for High-Resolution Large Vision-Language Modelsβ33Updated 3 months ago
- Code for Math-LLaVA: Bootstrapping Mathematical Reasoning for Multimodal Large Language Modelsβ91Updated last year
- [ICML 2025] M-STAR (Multimodal Self-Evolving TrAining for Reasoning) Project. Diving into Self-Evolving Training for Multimodal Reasoningβ69Updated 3 months ago