lhanchao777 / LVLM-Hallucinations-SurveyView external linksLinks
This is the first released survey paper on hallucinations of large vision-language models (LVLMs). To keep track of this field and continuously update our survey, we maintain this repository of relevant references.
β91Jul 26, 2024Updated last year
Alternatives and similar repositories for LVLM-Hallucinations-Survey
Users that are interested in LVLM-Hallucinations-Survey are comparing it to the libraries listed below
Sorting:
- π A curated list of resources dedicated to hallucination of multimodal large language models (MLLM).β979Sep 27, 2025Updated 4 months ago
- An LLM-free Multi-dimensional Benchmark for Multi-modal Hallucination Evaluationβ153Jan 15, 2024Updated 2 years ago
- The official GitHub page for ''Evaluating Object Hallucination in Large Vision-Language Models''β247Aug 21, 2025Updated 5 months ago
- An automatic MLLM hallucination detection frameworkβ19Sep 26, 2023Updated 2 years ago
- Less is More: Mitigating Multimodal Hallucination from an EOS Decision Perspective (ACL 2024)β57Oct 28, 2024Updated last year
- Github repository for "Why Is Spatial Reasoning Hard for VLMs? An Attention Mechanism Perspective on Focus Areas" (ICML 2025)β68May 2, 2025Updated 9 months ago
- β27Apr 18, 2025Updated 10 months ago
- [ICLR 2025] MLLM can see? Dynamic Correction Decoding for Hallucination Mitigationβ133Sep 11, 2025Updated 5 months ago
- Official pytorch implementation of "RITUAL: Random Image Transformations as a Universal Anti-hallucination Lever in Large Vision Languageβ¦β14Dec 16, 2024Updated last year
- [ICLR 2024] Analyzing and Mitigating Object Hallucination in Large Vision-Language Modelsβ155Apr 30, 2024Updated last year
- [CVPR 2025] Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local Attentionβ61Jul 16, 2024Updated last year
- [CVPR 2024 Highlight] Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decodingβ378Oct 7, 2024Updated last year
- HallE-Control: Controlling Object Hallucination in LMMsβ31Apr 10, 2024Updated last year
- [ACM Multimedia 2025] This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visualβ¦β82Feb 22, 2025Updated 11 months ago
- [NeurIPS 2025] Official Implementation for "Enhancing Vision-Language Model Reliability with Uncertainty-Guided Dropout Decoding"β22Dec 8, 2024Updated last year
- The First to Know: How Token Distributions Reveal Hidden Knowledge in Large Vision-Language Models?β42Nov 1, 2024Updated last year
- β13Feb 1, 2022Updated 4 years ago
- π curated list of awesome LMM hallucinations papers, methods & resources.β150Mar 23, 2024Updated last year
- [AAAI 26 Demo] Offical repo for CAT-V - Caption Anything in Video: Object-centric Dense Video Captioning with Spatiotemporal Multimodal Pβ¦β64Jan 27, 2026Updated 3 weeks ago
- Synthetic data generation for evaluating LLM symbolic and logic reasoningβ22Mar 20, 2025Updated 10 months ago
- [ACL 2025 Findings] Official pytorch implementation of "Don't Miss the Forest for the Trees: Attentional Vision Calibration for Large Visβ¦β24Jul 21, 2024Updated last year
- [CVPR2024 Highlight] Official implementation for Transferable Visual Prompting. The paper "Exploring the Transferability of Visual Promptβ¦β46Dec 20, 2024Updated last year
- [ECCV2024] Reflective Instruction Tuning: Mitigating Hallucinations in Large Vision-Language Modelsβ20Jul 17, 2024Updated last year
- up-to-date curated list of state-of-the-art Large vision language models hallucinations research work, papers & resourcesβ265Feb 8, 2026Updated last week
- LLM hallucination paper listβ331Mar 11, 2024Updated last year
- [ICLR '25] Official Pytorch implementation of "Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations"β95Nov 30, 2025Updated 2 months ago
- the official repo for EMNLP 2024 (main) paper "EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimoβ¦β20Apr 9, 2025Updated 10 months ago
- Can I Trust Your Answer? Visually Grounded Video Question Answering (CVPR'24, Highlight)β83Jul 1, 2024Updated last year
- The implementation of paper "EliMRec: Eliminating single-modal bias in multimedia recommendation", MM'22.β22Dec 7, 2023Updated 2 years ago
- Implementation of our IJCAI2022 oral paper, ER-SAN: Enhanced-Adaptive Relation Self-Attention Network for Image Captioning.β24Aug 5, 2023Updated 2 years ago
- [NeurIPS 2024] Calibrated Self-Rewarding Vision Language Modelsβ85Oct 26, 2025Updated 3 months ago
- Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimizationβ100Jan 30, 2024Updated 2 years ago
- [Arxiv] Aligning Modalities in Vision Large Language Models via Preference Fine-tuningβ91Apr 30, 2024Updated last year
- mPLUG-HalOwl: Multimodal Hallucination Evaluation and Mitigatingβ97Jan 29, 2024Updated 2 years ago
- Chinese Vision-Language Understanding Evaluationβ23Dec 26, 2024Updated last year
- [CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(β¦β325Oct 14, 2025Updated 4 months ago
- β55Apr 1, 2024Updated last year
- Official Implementation (Pytorch) of the "VidChain: Chain-of-Tasks with Metric-based Direct Preference Optimization for Dense Video Captiβ¦β23Jan 26, 2025Updated last year
- [ICML 2024] Official implementation for "HALC: Object Hallucination Reduction via Adaptive Focal-Contrast Decoding"β108Dec 4, 2024Updated last year