Theia-4869 / VisPrunerLinks
[ICCV 2025] Official code for paper: Beyond Text-Visual Attention: Exploiting Visual Cues for Effective Token Pruning in VLMs
☆25Updated 2 months ago
Alternatives and similar repositories for VisPruner
Users that are interested in VisPruner are comparing it to the libraries listed below
Sorting:
- [CVPR 2025] DyCoke: Dynamic Compression of Tokens for Fast Video Large Language Models☆72Updated 2 months ago
- Official code for paper: [CLS] Attention is All You Need for Training-Free Visual Token Pruning: Make VLM Inference Faster.☆86Updated 2 months ago
- 🚀 Video Compression Commander: Plug-and-Play Inference Acceleration for Video Large Language Models☆29Updated 2 months ago
- [CVPR 2025] PVC: Progressive Visual Token Compression for Unified Image and Video Processing in Large Vision-Language Models☆46Updated 2 months ago
- (CVPR 2025) PyramidDrop: Accelerating Your Large Vision-Language Models via Pyramid Visual Redundancy Reduction☆120Updated 5 months ago
- Official implementation of "Traceable Evidence Enhanced Visual Grounded Reasoning: Evaluation and Methodology"☆60Updated last month
- ☆54Updated 3 months ago
- ☆25Updated 6 months ago
- [AAAI 2025] HiRED strategically drops visual tokens in the image encoding stage to improve inference efficiency for High-Resolution Visio…☆41Updated 4 months ago
- ☆105Updated 5 months ago
- Official Repository: A Comprehensive Benchmark for Logical Reasoning in MLLMs☆40Updated 2 months ago
- LLaVA-PruMerge: Adaptive Token Reduction for Efficient Large Multimodal Models☆142Updated 2 months ago
- VeriThinker: Learning to Verify Makes Reasoning Model Efficient☆52Updated last month
- [CVPR 2025] DivPrune: Diversity-based Visual Token Pruning for Large Multimodal Models☆42Updated 3 months ago
- Dimple, the first Discrete Diffusion Multimodal Large Language Model☆95Updated last month
- Survey: https://arxiv.org/pdf/2507.20198☆121Updated this week
- A paper list about Token Merge, Reduce, Resample, Drop for MLLMs.☆69Updated 7 months ago
- [ICML 2024] CrossGET: Cross-Guided Ensemble of Tokens for Accelerating Vision-Language Transformers.☆34Updated 8 months ago
- Adapting LLaMA Decoder to Vision Transformer☆30Updated last year
- Autoregressive Semantic Visual Reconstruction Helps VLMs Understand Better☆37Updated 2 months ago
- [ICLR 2025] See What You Are Told: Visual Attention Sink in Large Multimodal Models☆45Updated 6 months ago
- [EMNLP 2024] Official code for "Beyond Embeddings: The Promise of Visual Table in Multi-Modal Models"☆20Updated 10 months ago
- [ICLR2025] γ -MOD: Mixture-of-Depth Adaptation for Multimodal Large Language Models☆38Updated 6 months ago
- Official code for paper "GRIT: Teaching MLLMs to Think with Images"☆121Updated 3 weeks ago
- HoliTom: Holistic Token Merging for Fast Video Large Language Models☆39Updated 2 weeks ago
- MADTP: Multimodal Alignment-Guided Dynamic Token Pruning for Accelerating Vision-Language Transformer☆46Updated 11 months ago
- [ICCV 2025] Dynamic-VLM☆24Updated 8 months ago
- [CVPR 2025] Mono-InternVL: Pushing the Boundaries of Monolithic Multimodal Large Language Models with Endogenous Visual Pre-training☆80Updated last month
- [ICCV'25] The official code of paper "Combining Similarity and Importance for Video Token Reduction on Large Visual Language Models"☆56Updated this week
- VCR-Bench: A Comprehensive Evaluation Framework for Video Chain-of-Thought Reasoning☆32Updated last month