shuyansy / MLLM-Semantic-HallucinationLinks
Exploring and mitigating semantic hallucinations in scene text perception and reasoning
☆11Updated last month
Alternatives and similar repositories for MLLM-Semantic-Hallucination
Users that are interested in MLLM-Semantic-Hallucination are comparing it to the libraries listed below
Sorting:
- Repo of FocusedAD☆13Updated 3 months ago
- Code for paper: Unified Text-to-Image Generation and Retrieval☆15Updated last year
- "Towards Improving Document Understanding: An Exploration on Text-Grounding via MLLMs" 2023☆14Updated 7 months ago
- [CVPR2025] VDocRAG: Retirval-Augmented Generation over Visually-Rich Documents☆31Updated last month
- [ECCV'24 Oral] PiTe: Pixel-Temporal Alignment for Large Video-Language Model☆16Updated 5 months ago
- ZoomEye: Enhancing Multimodal LLMs with Human-Like Zooming Capabilities through Tree-Based Image Exploration☆46Updated 6 months ago
- ☆42Updated 8 months ago
- The official implementation of the paper "MMFuser: Multimodal Multi-Layer Feature Fuser for Fine-Grained Vision-Language Understanding". …☆56Updated 8 months ago
- Code and data for paper "Exploring Hallucination of Large Multimodal Models in Video Understanding: Benchmark, Analysis and Mitigation".☆16Updated 2 months ago
- LMM solved catastrophic forgetting, AAAI2025☆44Updated 3 months ago
- Fast-Slow Thinking for Large Vision-Language Model Reasoning☆16Updated 2 months ago
- Code for Let LLMs Break Free from Overthinking via Self-Braking Tuning. https://arxiv.org/abs/2505.14604☆43Updated last month
- PyTorch implementation of "UNIT: Unifying Image and Text Recognition in One Vision Encoder", NeurlPS 2024.☆30Updated 9 months ago
- [EMNLP 2024] Official code for "Beyond Embeddings: The Promise of Visual Table in Multi-Modal Models"☆20Updated 9 months ago
- iLLaVA: An Image is Worth Fewer Than 1/3 Input Tokens in Large Multimodal Models☆19Updated 5 months ago
- Official InfiniBench: A Benchmark for Large Multi-Modal Models in Long-Form Movies and TV Shows☆15Updated last month
- Code for our Paper "All in an Aggregated Image for In-Image Learning"☆30Updated last year
- (ICLR2025 Spotlight) DEEM: Official implementation of Diffusion models serve as the eyes of large language models for image perception.☆35Updated 2 weeks ago
- [ICLR2025] MMIU: Multimodal Multi-image Understanding for Evaluating Large Vision-Language Models☆81Updated 10 months ago
- Implementation for "The Scalability of Simplicity: Empirical Analysis of Vision-Language Learning with a Single Transformer"☆51Updated 2 weeks ago
- ☆11Updated 3 months ago
- [NeurIPS 2024] TransAgent: Transfer Vision-Language Foundation Models with Heterogeneous Agent Collaboration☆24Updated 9 months ago
- Official Repository of Personalized Visual Instruct Tuning☆31Updated 4 months ago
- [CVPR 2025] Few-shot Recognition via Stage-Wise Retrieval-Augmented Finetuning☆20Updated 3 weeks ago
- Official implement of MIA-DPO☆59Updated 5 months ago
- ☆29Updated 2 years ago
- ☆48Updated 3 months ago
- [NeurIPS-24] This is the official implementation of the paper "DeepStack: Deeply Stacking Visual Tokens is Surprisingly Simple and Effect…☆37Updated last year
- WorldSense: Evaluating Real-world Omnimodal Understanding for Multimodal LLMs☆26Updated 2 months ago
- ☆23Updated last year