RainBowLuoCS / MMEvolLinks
(ACL 2025) π₯π₯π₯Code for "Empowering Multimodal Large Language Models with Evol-Instruct"
β16Updated last month
Alternatives and similar repositories for MMEvol
Users that are interested in MMEvol are comparing it to the libraries listed below
Sorting:
- [COLING 2024 (Oral)] PromISe:Releasing the Capabilities of LLMs with Prompt Introspective Searchβ21Updated 10 months ago
- Official resource for paper Investigating and Mitigating the Multimodal Hallucination Snowballing in Large Vision-Language Models (ACL 20β¦β12Updated 10 months ago
- (ICLR2025 Spotlight) DEEM: Official implementation of Diffusion models serve as the eyes of large language models for image perception.β34Updated 3 months ago
- PyTorch Implementation of "Divide, Conquer and Combine: A Training-Free Framework for High-Resolution Image Perception in Multimodal Largβ¦β23Updated last month
- Instruction Tuning in Continual Learning paradigmβ50Updated 4 months ago
- The First to Know: How Token Distributions Reveal Hidden Knowledge in Large Vision-Language Models?β31Updated 7 months ago
- [ICLR 2025] MLLM can see? Dynamic Correction Decoding for Hallucination Mitigationβ86Updated 6 months ago
- β74Updated last year
- [NeurIPS 2024] Calibrated Self-Rewarding Vision Language Modelsβ76Updated last year
- [EMNLP'23] The official GitHub page for ''Evaluating Object Hallucination in Large Vision-Language Models''β84Updated last year
- [ICML 2024] Official implementation for "HALC: Object Hallucination Reduction via Adaptive Focal-Contrast Decoding"β88Updated 6 months ago
- Official Code and data for ACL 2024 finding, "An Empirical Study on Parameter-Efficient Fine-Tuning for MultiModal Large Language Models"β19Updated 7 months ago
- VoCoT: Unleashing Visually Grounded Multi-Step Reasoning in Large Multi-Modal Modelsβ65Updated 11 months ago
- [CVPR' 25] Interleaved-Modal Chain-of-Thoughtβ53Updated 2 months ago
- Code for ICLR 2025 Paper: Visual Description Grounding Reduces Hallucinations and Boosts Reasoning in LVLMsβ16Updated last month
- β37Updated 11 months ago
- More Thinking, Less Seeing? Assessing Amplified Hallucination in Multimodal Reasoning Modelsβ21Updated 3 weeks ago
- Papers about Hallucination in Multi-Modal Large Language Models (MLLMs)β91Updated 7 months ago
- [EMNLP 2024] mDPO: Conditional Preference Optimization for Multimodal Large Language Models.β75Updated 7 months ago
- β86Updated 3 months ago
- HallE-Control: Controlling Object Hallucination in LMMsβ31Updated last year
- [CVPR 2025] Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local Attentionβ35Updated 11 months ago
- HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction Data (Accepted by CVPR 2024)β45Updated 11 months ago
- [CVPR 2025] Devils in Middle Layers of Large Vision-Language Models: Interpreting, Detecting and Mitigating Object Hallucinations via Attβ¦β23Updated 3 months ago
- Less is More: Mitigating Multimodal Hallucination from an EOS Decision Perspective (ACL 2024)β51Updated 7 months ago
- β122Updated 4 months ago
- This repository will continuously update the latest papers, technical reports, benchmarks about multimodal reasoning!β45Updated 3 months ago
- β47Updated 7 months ago
- Code for paper: Nullu: Mitigating Object Hallucinations in Large Vision-Language Models via HalluSpace Projectionβ31Updated 3 months ago
- Imagine While Reasoning in Space: Multimodal Visualization-of-Thought (ICML 2025)β27Updated 2 months ago