lhanchao777 / LVLM-Hallucinations-SurveyLinks
This is the first released survey paper on hallucinations of large vision-language models (LVLMs). To keep track of this field and continuously update our survey, we maintain this repository of relevant references.
β71Updated 10 months ago
Alternatives and similar repositories for LVLM-Hallucinations-Survey
Users that are interested in LVLM-Hallucinations-Survey are comparing it to the libraries listed below
Sorting:
- [ICML 2024] Official implementation for "HALC: Object Hallucination Reduction via Adaptive Focal-Contrast Decoding"β87Updated 6 months ago
- π curated list of awesome LMM hallucinations papers, methods & resources.β149Updated last year
- CHAIR metric is a rule-based metric for evaluating object hallucination in caption generation.β29Updated last year
- up-to-date curated list of state-of-the-art Large vision language models hallucinations research work, papers & resourcesβ138Updated last month
- Papers about Hallucination in Multi-Modal Large Language Models (MLLMs)β90Updated 7 months ago
- mPLUG-HalOwl: Multimodal Hallucination Evaluation and Mitigatingβ95Updated last year
- [ICLR 2024] Analyzing and Mitigating Object Hallucination in Large Vision-Language Modelsβ145Updated last year
- [ICLR 2025] MLLM can see? Dynamic Correction Decoding for Hallucination Mitigationβ86Updated 6 months ago
- [EMNLP 2024 Findings] The official PyTorch implementation of EchoSight: Advancing Visual-Language Models with Wiki Knowledge.β62Updated 3 months ago
- [CVPR 2024 Highlight] Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decodingβ285Updated 8 months ago
- [CVPR' 25] Interleaved-Modal Chain-of-Thoughtβ52Updated 2 months ago
- β46Updated 6 months ago
- β121Updated 4 months ago
- β74Updated last year
- The official GitHub page for ''Evaluating Object Hallucination in Large Vision-Language Models''β207Updated last year
- Official resource for paper Investigating and Mitigating the Multimodal Hallucination Snowballing in Large Vision-Language Models (ACL 20β¦β12Updated 10 months ago
- MMICL, a state-of-the-art VLM with the in context learning ability from ICL, PKUβ47Updated last year
- β101Updated 2 months ago
- A Comprehensive Survey on Evaluating Reasoning Capabilities in Multimodal Large Language Models.β64Updated 3 months ago
- This repository will continuously update the latest papers, technical reports, benchmarks about multimodal reasoning!β44Updated 3 months ago
- [CVPR25] A ChatGPT-Prompted Visual hallucination Evaluation Dataset, featuring over 100,000 data samples and four advanced evaluation modβ¦β17Updated 2 months ago
- [ICLR'25] Official code for the paper 'MLLMs Know Where to Look: Training-free Perception of Small Visual Details with Multimodal LLMs'β212Updated 2 months ago
- [ECCV 2024] Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMsβ123Updated 7 months ago
- β9Updated last year
- VoCoT: Unleashing Visually Grounded Multi-Step Reasoning in Large Multi-Modal Modelsβ65Updated 11 months ago
- [NeurIPS 2024] Mitigating Object Hallucination via Concentric Causal Attentionβ56Updated 5 months ago
- Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimizationβ88Updated last year
- This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding stratβ¦β78Updated 4 months ago
- Code for Reducing Hallucinations in Vision-Language Models via Latent Space Steeringβ58Updated 7 months ago
- Latest Advances on Modality Priors in Multimodal Large Language Modelsβ20Updated last month